[ 470.616067] env[61964]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61964) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.616398] env[61964]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61964) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.616511] env[61964]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61964) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.616801] env[61964]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 470.711992] env[61964]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61964) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 470.721823] env[61964]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61964) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 471.325689] env[61964]: INFO nova.virt.driver [None req-b283eae0-7cc0-44ed-af87-8ae28b4dd2f9 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 471.397782] env[61964]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.397967] env[61964]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.398067] env[61964]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61964) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 474.596991] env[61964]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-a3c7dd23-d008-45a3-9718-27322483fcc0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.614124] env[61964]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61964) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 474.614276] env[61964]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-9d8ad1ef-bcc9-4257-99f2-6cca4952d2d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.648999] env[61964]: INFO oslo_vmware.api [-] Successfully established new session; session ID is c00f9. [ 474.649177] env[61964]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.251s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.649727] env[61964]: INFO nova.virt.vmwareapi.driver [None req-b283eae0-7cc0-44ed-af87-8ae28b4dd2f9 None None] VMware vCenter version: 7.0.3 [ 474.653165] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c1680b-3659-4caa-a133-686ec103f368 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.671215] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cc524e-22bd-4158-a944-f43b9c5465ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.677301] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c781de-115e-4d2c-80c0-70e57b5e726e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.683954] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1860ebaa-4d9b-4ad8-bbf1-a138c7a75ac5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.697422] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adeeaab4-b8ea-4a2b-b8b4-758bc42d7b35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.703569] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055a49e5-48ad-435e-b894-d71465a8ba58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.734546] env[61964]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-c70f9db3-5380-4e0a-bd04-61361126a045 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.739531] env[61964]: DEBUG nova.virt.vmwareapi.driver [None req-b283eae0-7cc0-44ed-af87-8ae28b4dd2f9 None None] Extension org.openstack.compute already exists. {{(pid=61964) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 474.742441] env[61964]: INFO nova.compute.provider_config [None req-b283eae0-7cc0-44ed-af87-8ae28b4dd2f9 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 475.246330] env[61964]: DEBUG nova.context [None req-b283eae0-7cc0-44ed-af87-8ae28b4dd2f9 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),ba77e97d-6450-4fd4-9b37-f76f338f412c(cell1) {{(pid=61964) load_cells /opt/stack/nova/nova/context.py:464}} [ 475.248389] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.248621] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.249314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.249746] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Acquiring lock "ba77e97d-6450-4fd4-9b37-f76f338f412c" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.249932] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Lock "ba77e97d-6450-4fd4-9b37-f76f338f412c" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.251108] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Lock "ba77e97d-6450-4fd4-9b37-f76f338f412c" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.272597] env[61964]: INFO dbcounter [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Registered counter for database nova_cell0 [ 475.280647] env[61964]: INFO dbcounter [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Registered counter for database nova_cell1 [ 475.283928] env[61964]: DEBUG oslo_db.sqlalchemy.engines [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61964) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.284248] env[61964]: DEBUG oslo_db.sqlalchemy.engines [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61964) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.289010] env[61964]: ERROR nova.db.main.api [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.289010] env[61964]: result = function(*args, **kwargs) [ 475.289010] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.289010] env[61964]: return func(*args, **kwargs) [ 475.289010] env[61964]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.289010] env[61964]: result = fn(*args, **kwargs) [ 475.289010] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.289010] env[61964]: return f(*args, **kwargs) [ 475.289010] env[61964]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.289010] env[61964]: return db.service_get_minimum_version(context, binaries) [ 475.289010] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.289010] env[61964]: _check_db_access() [ 475.289010] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.289010] env[61964]: stacktrace = ''.join(traceback.format_stack()) [ 475.289010] env[61964]: [ 475.290028] env[61964]: ERROR nova.db.main.api [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.290028] env[61964]: result = function(*args, **kwargs) [ 475.290028] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.290028] env[61964]: return func(*args, **kwargs) [ 475.290028] env[61964]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.290028] env[61964]: result = fn(*args, **kwargs) [ 475.290028] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.290028] env[61964]: return f(*args, **kwargs) [ 475.290028] env[61964]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.290028] env[61964]: return db.service_get_minimum_version(context, binaries) [ 475.290028] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.290028] env[61964]: _check_db_access() [ 475.290028] env[61964]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.290028] env[61964]: stacktrace = ''.join(traceback.format_stack()) [ 475.290028] env[61964]: [ 475.290473] env[61964]: WARNING nova.objects.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 475.290515] env[61964]: WARNING nova.objects.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Failed to get minimum service version for cell ba77e97d-6450-4fd4-9b37-f76f338f412c [ 475.290968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Acquiring lock "singleton_lock" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 475.291146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Acquired lock "singleton_lock" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 475.291392] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Releasing lock "singleton_lock" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 475.291710] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Full set of CONF: {{(pid=61964) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 475.291858] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ******************************************************************************** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 475.291983] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Configuration options gathered from: {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 475.292160] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 475.292354] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 475.292480] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ================================================================================ {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 475.292690] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] allow_resize_to_same_host = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.292861] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] arq_binding_timeout = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.292993] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] backdoor_port = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.293158] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] backdoor_socket = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.293338] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] block_device_allocate_retries = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.293506] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] block_device_allocate_retries_interval = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.293677] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cert = self.pem {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.293843] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294052] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute_monitors = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294199] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] config_dir = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294376] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] config_drive_format = iso9660 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294536] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294713] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] config_source = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.294888] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] console_host = devstack {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295062] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] control_exchange = nova {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295224] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cpu_allocation_ratio = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295383] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] daemon = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295549] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] debug = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295703] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_access_ip_network_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.295865] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_availability_zone = nova {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296030] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_ephemeral_format = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296194] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_green_pool_size = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296433] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296598] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] default_schedule_zone = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296777] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] disk_allocation_ratio = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.296950] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] enable_new_services = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297146] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] enabled_apis = ['osapi_compute'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297314] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] enabled_ssl_apis = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297476] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] flat_injected = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297634] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] force_config_drive = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297791] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] force_raw_images = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.297958] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] graceful_shutdown_timeout = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.298131] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] heal_instance_info_cache_interval = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.298347] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] host = cpu-1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.298521] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.298683] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.298843] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.299068] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.299241] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_build_timeout = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.299404] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_delete_interval = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.299571] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_format = [instance: %(uuid)s] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.299763] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_name_template = instance-%08x {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300009] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_usage_audit = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300184] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_usage_audit_period = month {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300356] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300522] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300685] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] internal_service_availability_zone = internal {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300841] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] key = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.300997] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] live_migration_retry_count = 30 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301180] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_color = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301350] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_config_append = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301516] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301678] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_dir = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301836] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.301966] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_options = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.302175] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_rotate_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.302373] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_rotate_interval_type = days {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.302579] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] log_rotation_type = none {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.302746] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.302962] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.303212] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.303416] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.303577] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.303776] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] long_rpc_timeout = 1800 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.303974] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_concurrent_builds = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.304195] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_concurrent_live_migrations = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.304396] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_concurrent_snapshots = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.304589] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_local_block_devices = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.304785] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_logfile_count = 30 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.304983] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] max_logfile_size_mb = 200 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.305205] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] maximum_instance_delete_attempts = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.305412] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metadata_listen = 0.0.0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.305614] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metadata_listen_port = 8775 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.305817] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metadata_workers = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306066] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] migrate_max_retries = -1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306269] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] mkisofs_cmd = genisoimage {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306479] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306614] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] my_ip = 10.180.1.21 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306820] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.306987] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] network_allocate_retries = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.307183] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.307358] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.307521] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] osapi_compute_listen_port = 8774 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.307688] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] osapi_compute_unique_server_name_scope = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.307858] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] osapi_compute_workers = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308031] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] password_length = 12 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308197] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] periodic_enable = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308361] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] periodic_fuzzy_delay = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308530] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] pointer_model = usbtablet {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308695] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] preallocate_images = none {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308855] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] publish_errors = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.308984] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] pybasedir = /opt/stack/nova {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309158] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ram_allocation_ratio = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309323] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rate_limit_burst = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309486] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rate_limit_except_level = CRITICAL {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309645] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rate_limit_interval = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309805] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reboot_timeout = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.309963] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reclaim_instance_interval = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310133] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] record = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310302] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reimage_timeout_per_gb = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310467] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] report_interval = 120 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310629] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rescue_timeout = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310790] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reserved_host_cpus = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.310947] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reserved_host_disk_mb = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311118] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reserved_host_memory_mb = 512 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311281] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] reserved_huge_pages = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311442] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] resize_confirm_window = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311601] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] resize_fs_using_block_device = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311761] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] resume_guests_state_on_host_boot = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.311963] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312181] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] rpc_response_timeout = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312352] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] run_external_periodic_tasks = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312522] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] running_deleted_instance_action = reap {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312683] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312841] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] running_deleted_instance_timeout = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.312999] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler_instance_sync_interval = 120 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313211] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_down_time = 720 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313352] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] servicegroup_driver = db {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313510] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] shell_completion = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313673] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] shelved_offload_time = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313829] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] shelved_poll_interval = 3600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.313993] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] shutdown_timeout = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.314168] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] source_is_ipv6 = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.314331] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ssl_only = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.314599] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.314772] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] sync_power_state_interval = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.314937] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] sync_power_state_pool_size = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315228] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] syslog_log_facility = LOG_USER {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315284] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] tempdir = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315448] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] timeout_nbd = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315612] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] transport_url = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315791] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] update_resources_interval = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.315938] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_cow_images = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316109] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_eventlog = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316273] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_journal = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316432] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_json = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316589] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_rootwrap_daemon = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316747] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_stderr = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.316905] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] use_syslog = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317074] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vcpu_pin_set = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317243] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plugging_is_fatal = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317414] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plugging_timeout = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317578] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] virt_mkfs = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317740] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] volume_usage_poll_interval = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.317917] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] watch_log_file = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.318113] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] web = /usr/share/spice-html5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.318301] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_concurrency.disable_process_locking = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.318582] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.318762] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.318928] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319112] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319289] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319456] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319635] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.auth_strategy = keystone {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319802] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.compute_link_prefix = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.319981] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.320201] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.dhcp_domain = novalocal {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.320388] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.enable_instance_password = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.320555] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.glance_link_prefix = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.320724] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.320899] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321075] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.instance_list_per_project_cells = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321239] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.list_records_by_skipping_down_cells = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321402] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.local_metadata_per_cell = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321570] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.max_limit = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321739] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.metadata_cache_expiration = 15 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.321911] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.neutron_default_tenant_id = default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322141] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.response_validation = warn {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322295] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.use_neutron_default_nets = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322471] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322639] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322807] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.322982] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.323177] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_dynamic_targets = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.323360] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_jsonfile_path = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.323547] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.323738] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.backend = dogpile.cache.memcached {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.323941] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.backend_argument = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324141] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.config_prefix = cache.oslo {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324317] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.dead_timeout = 60.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324484] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.debug_cache_backend = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324643] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.enable_retry_client = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324801] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.enable_socket_keepalive = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.324970] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.enabled = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325148] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.enforce_fips_mode = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325315] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.expiration_time = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325481] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.hashclient_retry_attempts = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325646] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325811] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_dead_retry = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.325995] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_password = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.326187] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.326353] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.326524] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_pool_maxsize = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.326677] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.326837] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_sasl_enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327027] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327200] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327364] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.memcache_username = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327530] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.proxies = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327693] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_db = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.327853] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_password = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328033] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328212] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328382] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_server = localhost:6379 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328546] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_socket_timeout = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328707] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.redis_username = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.328871] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.retry_attempts = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329045] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.retry_delay = 0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329212] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.socket_keepalive_count = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329375] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.socket_keepalive_idle = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329537] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.socket_keepalive_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329697] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.tls_allowed_ciphers = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.329854] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.tls_cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330065] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.tls_certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330249] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.tls_enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330412] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cache.tls_keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330581] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330756] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.auth_type = password {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.330918] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331136] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331268] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331434] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331597] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.cross_az_attach = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331758] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.debug = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.331918] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.endpoint_template = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332123] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.http_retries = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332281] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332445] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332617] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.os_region_name = RegionOne {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332783] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.332943] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cinder.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333131] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333298] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.cpu_dedicated_set = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333458] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.cpu_shared_set = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333623] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.image_type_exclude_list = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333783] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.333945] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334121] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334284] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334452] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334617] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.resource_provider_association_refresh = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334777] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.334939] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.shutdown_retry_interval = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335130] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335309] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] conductor.workers = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335484] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] console.allowed_origins = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335644] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] console.ssl_ciphers = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335814] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] console.ssl_minimum_version = default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.335979] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] consoleauth.enforce_session_timeout = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336161] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] consoleauth.token_ttl = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336334] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336494] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336658] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336818] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.336977] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337149] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337313] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337470] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337628] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337786] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.337944] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338115] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338278] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338446] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.service_type = accelerator {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338606] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338767] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.338924] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339089] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339272] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339434] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] cyborg.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339610] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.backend = sqlalchemy {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339779] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.connection = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.339946] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.connection_debug = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340149] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.connection_parameters = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340325] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.connection_recycle_time = 3600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340487] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.connection_trace = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340648] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.db_inc_retry_interval = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340811] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.db_max_retries = 20 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.340974] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.db_max_retry_interval = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341149] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.db_retry_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341316] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.max_overflow = 50 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341478] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.max_pool_size = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341637] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.max_retries = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341806] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.341964] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.mysql_wsrep_sync_wait = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.342165] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.pool_timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.342337] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.retry_interval = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.342501] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.slave_connection = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.342661] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.sqlite_synchronous = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.342823] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] database.use_db_reconnect = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343008] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.backend = sqlalchemy {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343213] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.connection = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343382] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.connection_debug = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343554] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.connection_parameters = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343720] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.connection_recycle_time = 3600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.343884] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.connection_trace = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344055] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.db_inc_retry_interval = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344224] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.db_max_retries = 20 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344387] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.db_max_retry_interval = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344549] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.db_retry_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344712] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.max_overflow = 50 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.344871] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.max_pool_size = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345040] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.max_retries = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345213] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345374] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345532] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.pool_timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345693] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.retry_interval = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.345850] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.slave_connection = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346024] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] api_database.sqlite_synchronous = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346198] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] devices.enabled_mdev_types = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346377] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346548] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346710] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ephemeral_storage_encryption.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.346873] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347056] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.api_servers = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347232] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347391] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347553] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347710] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.347866] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348035] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.debug = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348204] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.default_trusted_certificate_ids = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348368] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.enable_certificate_validation = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348530] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.enable_rbd_download = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348689] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.348853] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349023] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349185] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349345] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349506] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.num_retries = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349673] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.rbd_ceph_conf = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.349835] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.rbd_connect_timeout = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350020] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.rbd_pool = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350218] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.rbd_user = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350385] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350548] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350708] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.350877] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.service_type = image {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351050] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351213] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351373] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351541] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351722] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.351888] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.verify_glance_signatures = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352058] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] glance.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352258] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] guestfs.debug = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352433] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352598] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352757] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.352916] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353099] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353280] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353442] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353600] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353760] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.353916] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354083] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354243] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354402] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354557] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354719] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.354888] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.service_type = shared-file-system {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355054] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.share_apply_policy_timeout = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355216] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355376] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355532] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355688] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.355868] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.356035] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] manila.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.356205] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] mks.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.356559] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.356751] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.manager_interval = 2400 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.356924] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.precache_concurrency = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357108] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.remove_unused_base_images = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357284] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357453] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357631] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] image_cache.subdirectory_name = _base {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357806] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.api_max_retries = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.357973] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.api_retry_interval = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358146] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358311] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358472] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358630] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358794] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.358958] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.conductor_group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359132] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359297] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359455] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359618] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359775] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.359934] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360128] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360305] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.peer_list = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360466] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360625] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360790] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.serial_console_state_timeout = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.360949] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361132] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.service_type = baremetal {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361295] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.shard = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361458] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361614] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361769] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.361924] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.362141] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.362309] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ironic.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.362494] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.362667] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] key_manager.fixed_key = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.362847] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363021] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.barbican_api_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363180] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.barbican_endpoint = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363352] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.barbican_endpoint_type = public {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363508] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.barbican_region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363662] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363818] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.363976] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364148] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364308] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364478] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.number_of_retries = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364634] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.retry_delay = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364792] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.send_service_user_token = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.364951] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365120] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365283] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.verify_ssl = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365441] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican.verify_ssl_path = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365604] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365765] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.365919] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366084] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366248] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366408] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366565] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366722] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.366879] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] barbican_service_user.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367055] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.approle_role_id = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367217] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.approle_secret_id = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367388] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.kv_mountpoint = secret {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367545] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.kv_path = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367709] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.kv_version = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.367868] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.namespace = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368034] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.root_token_id = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368196] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.ssl_ca_crt_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368362] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.timeout = 60.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368522] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.use_ssl = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368688] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.368855] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369025] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369187] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369347] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369510] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369666] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369819] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.369974] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370182] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370346] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370503] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370659] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370815] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.370972] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371141] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371311] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.service_type = identity {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371471] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371629] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371786] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.371946] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.372164] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.372382] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] keystone.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.372592] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.ceph_mount_options = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.373251] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.373452] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.connection_uri = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.373623] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_mode = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.373795] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.373968] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_models = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.374156] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_power_governor_high = performance {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.374333] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.374505] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_power_management = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.374673] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.374843] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.device_detach_attempts = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375042] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.device_detach_timeout = 20 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375188] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.disk_cachemodes = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375354] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.disk_prefix = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375520] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.enabled_perf_events = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375691] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.file_backed_memory = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.375862] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.gid_maps = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376036] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.hw_disk_discard = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376201] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.hw_machine_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376374] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_rbd_ceph_conf = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376541] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376707] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.376877] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_rbd_glance_store_name = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377058] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_rbd_pool = rbd {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377234] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_type = default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377396] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.images_volume_group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377558] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.inject_key = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377723] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.inject_partition = -2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.377885] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.inject_password = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378058] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.iscsi_iface = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378225] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.iser_use_multipath = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378395] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378560] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378725] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_downtime = 500 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.378891] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379064] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379229] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_inbound_addr = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379393] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379552] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379732] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_scheme = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.379887] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_timeout_action = abort {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380113] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_tunnelled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380262] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_uri = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380431] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.live_migration_with_native_tls = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380593] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.max_queues = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380757] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.380984] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.381163] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.nfs_mount_options = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.381447] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.381654] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.381784] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.381945] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.382151] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.382317] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_pcie_ports = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.382490] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.382660] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.pmem_namespaces = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.382819] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.quobyte_client_cfg = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.383111] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.383286] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.383455] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.383621] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.383782] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rbd_secret_uuid = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384112] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rbd_user = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384112] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384332] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384518] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rescue_image_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384777] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rescue_kernel_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384837] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rescue_ramdisk_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.384992] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.385167] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.rx_queue_size = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.385337] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.smbfs_mount_options = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.385610] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.385788] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.snapshot_compression = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.385943] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.snapshot_image_format = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.386192] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.386358] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.sparse_logical_volumes = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.386518] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.swtpm_enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.386687] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.swtpm_group = tss {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.386853] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.swtpm_user = tss {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387030] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.sysinfo_serial = unique {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387194] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.tb_cache_size = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387385] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.tx_queue_size = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387602] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.uid_maps = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387774] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.use_virtio_for_bridges = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.387947] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.virt_type = kvm {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388135] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.volume_clear = zero {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388307] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.volume_clear_size = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388472] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.volume_use_multipath = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388630] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_cache_path = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388796] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.388961] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.389139] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.389310] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.389586] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.389764] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.vzstorage_mount_user = stack {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.389933] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.390172] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.390371] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.auth_type = password {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.390536] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.390697] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.390858] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391024] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391188] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391361] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.default_floating_pool = public {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391518] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391681] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.extension_sync_interval = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.391843] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.http_retries = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392006] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392193] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392361] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392531] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392689] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.392854] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.ovs_bridge = br-int {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393031] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.physnets = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393195] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.region_name = RegionOne {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393354] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393564] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.service_metadata_proxy = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393678] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.393843] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.service_type = network {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394012] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394221] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394454] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394574] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394763] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.394920] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] neutron.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395104] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.bdms_in_notifications = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395290] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.default_level = INFO {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395461] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.include_share_mapping = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395636] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.notification_format = unversioned {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395798] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.notify_on_state_change = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.395973] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.396161] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] pci.alias = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.396338] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] pci.device_spec = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.396502] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] pci.report_in_placement = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.396674] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.396846] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.auth_type = password {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397020] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397184] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397342] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397503] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397661] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397818] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.397974] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.default_domain_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398143] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.default_domain_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398305] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.domain_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398459] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.domain_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398616] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398775] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.398930] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399096] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399257] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399452] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.password = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399627] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.project_domain_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399794] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.project_domain_name = Default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.399961] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.project_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.400170] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.project_name = service {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.400350] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.region_name = RegionOne {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.400565] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.400677] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.400844] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.service_type = placement {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401021] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401181] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401345] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401502] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.system_scope = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401659] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401822] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.trust_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.401973] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.user_domain_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.402173] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.user_domain_name = Default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.402338] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.user_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.402513] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.username = nova {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.402693] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.402870] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] placement.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403059] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.cores = 20 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403229] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.count_usage_from_placement = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403401] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403575] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.injected_file_content_bytes = 10240 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403740] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.injected_file_path_length = 255 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.403905] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.injected_files = 5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404084] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.instances = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404257] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.key_pairs = 100 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404434] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.metadata_items = 128 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404597] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.ram = 51200 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404759] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.recheck_quota = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.404925] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.server_group_members = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405104] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] quota.server_groups = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405281] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405449] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405605] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.image_metadata_prefilter = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405767] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.405932] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.max_attempts = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406107] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.max_placement_results = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406301] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406489] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406652] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406825] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] scheduler.workers = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.406999] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.407184] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.407366] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.407534] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.407697] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.407860] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408029] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408222] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408394] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.host_subset_size = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408559] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408719] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.408879] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409055] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.isolated_hosts = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409224] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.isolated_images = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409392] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409554] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409717] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.409880] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.pci_in_placement = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410073] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410288] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410463] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410640] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410803] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.410964] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411139] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.track_instance_changes = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411316] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411489] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metrics.required = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411651] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metrics.weight_multiplier = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411813] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.411991] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] metrics.weight_setting = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.412326] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.412507] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.412681] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.port_range = 10000:20000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.412881] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413034] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413207] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] serial_console.serialproxy_port = 6083 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413374] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413545] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.auth_type = password {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413703] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.413859] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414030] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414195] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414352] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414522] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.send_service_user_token = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414683] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.414838] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] service_user.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.415012] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.agent_enabled = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.415180] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.415487] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.415677] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.415845] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.html5proxy_port = 6082 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416014] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.image_compression = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416178] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.jpeg_compression = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416338] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.playback_compression = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416500] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.require_secure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416669] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.server_listen = 127.0.0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416838] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416998] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.streaming_mode = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417169] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] spice.zlib_compression = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417336] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] upgrade_levels.baseapi = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417506] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] upgrade_levels.compute = auto {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417666] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] upgrade_levels.conductor = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417823] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] upgrade_levels.scheduler = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417986] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418169] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418332] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418533] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418659] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418817] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418975] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419149] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419311] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vendordata_dynamic_auth.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419483] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.api_retry_count = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419644] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.ca_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419814] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419979] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.cluster_name = testcl1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420217] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.connection_pool_size = 10 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420390] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.console_delay_seconds = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420567] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.datastore_regex = ^datastore.* {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420774] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420946] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.host_password = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421127] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.host_port = 443 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421300] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.host_username = administrator@vsphere.local {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421467] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.insecure = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421629] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.integration_bridge = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421795] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.maximum_objects = 100 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421956] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.pbm_default_policy = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422160] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.pbm_enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422347] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.pbm_wsdl_location = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422525] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422688] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.serial_port_proxy_uri = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422847] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.serial_port_service_uri = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423021] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.task_poll_interval = 0.5 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423333] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.use_linked_clone = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423397] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.vnc_keymap = en-us {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423528] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.vnc_port = 5900 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423692] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vmware.vnc_port_total = 10000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423877] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.auth_schemes = ['none'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424065] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424404] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424599] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424774] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.novncproxy_port = 6080 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424953] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.server_listen = 127.0.0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425145] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425313] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.vencrypt_ca_certs = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425475] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.vencrypt_client_cert = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425637] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vnc.vencrypt_client_key = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425817] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425981] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_deep_image_inspection = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426160] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426327] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426490] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426662] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.disable_rootwrap = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426820] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.enable_numa_live_migration = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426984] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427160] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427329] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427494] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.libvirt_disable_apic = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427655] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427820] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427984] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428160] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428328] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428490] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428653] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428815] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428976] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429156] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429343] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429513] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.client_socket_timeout = 900 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429680] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.default_pool_size = 1000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429844] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.keep_alive = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430035] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.max_header_line = 16384 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430243] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430419] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.ssl_ca_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430593] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.ssl_cert_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430750] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.ssl_key_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430915] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.tcp_keepidle = 600 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431105] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431277] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] zvm.ca_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431437] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] zvm.cloud_connector_url = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431719] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431892] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] zvm.reachable_timeout = 300 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432081] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.enforce_new_defaults = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432492] env[61964]: WARNING oslo_config.cfg [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 475.432685] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.enforce_scope = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432867] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.policy_default_rule = default {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433064] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433251] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.policy_file = policy.yaml {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433429] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433595] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433758] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433921] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434097] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434298] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434490] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434670] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.connection_string = messaging:// {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434842] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.enabled = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435026] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.es_doc_type = notification {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435199] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.es_scroll_size = 10000 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435371] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.es_scroll_time = 2m {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435535] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.filter_error_trace = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435757] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.hmac_keys = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435877] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.sentinel_service_name = mymaster {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436035] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.socket_timeout = 0.1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436201] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.trace_requests = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436365] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler.trace_sqlalchemy = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436608] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler_jaeger.process_tags = {} {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436810] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler_jaeger.service_name_prefix = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436983] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] profiler_otlp.service_name_prefix = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437170] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] remote_debug.host = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437335] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] remote_debug.port = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437516] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437680] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437846] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438022] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438188] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438354] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438514] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438677] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438839] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439014] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439182] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439355] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439524] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439698] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439869] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440075] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440280] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440467] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440636] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440802] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440968] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441150] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441315] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441483] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441647] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441806] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441968] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442168] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442348] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442516] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442690] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442860] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443033] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443215] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443387] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443551] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443737] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443901] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_notifications.retry = -1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444097] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444276] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444448] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.auth_section = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444612] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.auth_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444770] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.cafile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444930] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.certfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445108] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.collect_timing = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445273] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.connect_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445432] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.connect_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445591] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.endpoint_id = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445749] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.endpoint_override = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445914] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.insecure = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446125] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.keyfile = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446310] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.max_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446473] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.min_version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446633] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.region_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446797] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.retriable_status_codes = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446960] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.service_name = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447134] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.service_type = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447302] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.split_loggers = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447487] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.status_code_retries = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447732] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.status_code_retry_delay = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447986] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.timeout = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448262] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.valid_interfaces = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448539] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_limit.version = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448822] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_reports.file_event_handler = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449104] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449374] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] oslo_reports.log_dir = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449579] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449749] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449913] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450115] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450299] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450462] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450634] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450795] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450955] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451133] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451305] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451465] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] vif_plug_ovs_privileged.user = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451638] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451819] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451993] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452204] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452393] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452561] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452730] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452895] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453089] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453269] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.isolate_vif = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453441] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453610] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453779] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453948] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454123] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_vif_ovs.per_port_bridge = False {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454294] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_brick.lock_path = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454460] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454624] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454795] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.capabilities = [21] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454955] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455127] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.helper_command = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455298] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455465] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455626] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] privsep_osbrick.user = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455799] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455977] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.group = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456132] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.helper_command = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456286] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456449] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456604] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] nova_sys_admin.user = None {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456734] env[61964]: DEBUG oslo_service.service [None req-cd6ec5c3-ae95-4b52-9fdd-440d94af0e20 None None] ******************************************************************************** {{(pid=61964) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 475.457226] env[61964]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 475.962023] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Getting list of instances from cluster (obj){ [ 475.962023] env[61964]: value = "domain-c8" [ 475.962023] env[61964]: _type = "ClusterComputeResource" [ 475.962023] env[61964]: } {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 475.962388] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886a5d66-6939-4961-b4d8-0c17880ecfc1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.972090] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Got total of 0 instances {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 475.972692] env[61964]: WARNING nova.virt.vmwareapi.driver [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 475.973166] env[61964]: INFO nova.virt.node [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Generated node identity 57b292ab-02d9-4aab-ba83-292890345a17 [ 475.973392] env[61964]: INFO nova.virt.node [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Wrote node identity 57b292ab-02d9-4aab-ba83-292890345a17 to /opt/stack/data/n-cpu-1/compute_id [ 476.475893] env[61964]: WARNING nova.compute.manager [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Compute nodes ['57b292ab-02d9-4aab-ba83-292890345a17'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 477.481514] env[61964]: INFO nova.compute.manager [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 478.486455] env[61964]: WARNING nova.compute.manager [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 478.486799] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.486919] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.487052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.487210] env[61964]: DEBUG nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 478.488137] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fecd8d6-2b69-4110-945c-88382763bf20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.496931] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc886a3-3edb-4c8a-8014-81f434b63c41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.512205] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a3b5c5-06e2-4828-9128-e7fc10e21697 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.519068] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbfd830-ef6c-422b-96d9-7f9a8fee1035 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.547492] env[61964]: DEBUG nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181607MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 478.547673] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.547821] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 479.049896] env[61964]: WARNING nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] No compute node record for cpu-1:57b292ab-02d9-4aab-ba83-292890345a17: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 57b292ab-02d9-4aab-ba83-292890345a17 could not be found. [ 479.553856] env[61964]: INFO nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 57b292ab-02d9-4aab-ba83-292890345a17 [ 481.062070] env[61964]: DEBUG nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 481.062441] env[61964]: DEBUG nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 481.213527] env[61964]: INFO nova.scheduler.client.report [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] [req-90ed36e7-ca62-425f-9407-9f6213175fd7] Created resource provider record via placement API for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 481.230407] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8304f134-4cf3-47d4-8161-0c173522c949 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.238860] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecefb45c-40cf-4a55-9994-ac2e7d6df515 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.268292] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dada88-52a5-4d11-a7c8-56c9c8849d98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.275212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271d0124-4637-4738-b9a5-25843a224129 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.287905] env[61964]: DEBUG nova.compute.provider_tree [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.823747] env[61964]: DEBUG nova.scheduler.client.report [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 481.824047] env[61964]: DEBUG nova.compute.provider_tree [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 0 to 1 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 481.824229] env[61964]: DEBUG nova.compute.provider_tree [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.877402] env[61964]: DEBUG nova.compute.provider_tree [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 1 to 2 during operation: update_traits {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 482.381869] env[61964]: DEBUG nova.compute.resource_tracker [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 482.382274] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.834s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 482.382324] env[61964]: DEBUG nova.service [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Creating RPC server for service compute {{(pid=61964) start /opt/stack/nova/nova/service.py:186}} [ 482.396241] env[61964]: DEBUG nova.service [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] Join ServiceGroup membership for this service compute {{(pid=61964) start /opt/stack/nova/nova/service.py:203}} [ 482.396466] env[61964]: DEBUG nova.servicegroup.drivers.db [None req-8c7b3afc-1545-4033-91dc-147ed3df194f None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61964) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 517.398169] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_power_states {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.400093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquiring lock "e23a7d59-0516-46b9-ab6d-e4c313ad758a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.400356] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "e23a7d59-0516-46b9-ab6d-e4c313ad758a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.902694] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Getting list of instances from cluster (obj){ [ 517.902694] env[61964]: value = "domain-c8" [ 517.902694] env[61964]: _type = "ClusterComputeResource" [ 517.902694] env[61964]: } {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 517.903965] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3562d151-c61d-4b09-ba39-d631fb85c759 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.913683] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 517.922289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Got total of 0 instances {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 517.922289] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.922482] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Getting list of instances from cluster (obj){ [ 517.922482] env[61964]: value = "domain-c8" [ 517.922482] env[61964]: _type = "ClusterComputeResource" [ 517.922482] env[61964]: } {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 517.923353] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df534da5-010a-4590-9521-4e983d4c050a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.935803] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Got total of 0 instances {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 518.467339] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.467598] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.470552] env[61964]: INFO nova.compute.claims [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.078059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "4e14a519-b543-4a18-9e1d-7df16abf5345" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.078866] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.276095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquiring lock "e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.276267] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.561854] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0563eca5-6818-456a-9de0-c2f0dfc439c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.571454] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1b87e-c0da-4208-8ae6-9bfec13f2c2e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.603036] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 519.610098] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271b46d4-786d-4a05-893a-864ab4135c48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.622343] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3855e641-40c4-4047-9dbe-51afaf739630 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.650147] env[61964]: DEBUG nova.compute.provider_tree [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.779331] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 520.154353] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.154353] env[61964]: DEBUG nova.scheduler.client.report [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 520.313681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.377104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "7ec4c677-9166-4b3c-ade7-548bee8e245b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.377818] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "7ec4c677-9166-4b3c-ade7-548bee8e245b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.661137] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.192s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.661137] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 520.663916] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.512s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.665478] env[61964]: INFO nova.compute.claims [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.881563] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 521.172678] env[61964]: DEBUG nova.compute.utils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.175863] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 521.175863] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 521.413723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.673701] env[61964]: DEBUG nova.policy [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2801bb2290240d39ec804d3e1027c52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b3dd060f27b4e81b46ee4c9c07ff132', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.688829] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 521.780216] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97a11e2-e20d-4deb-b21c-08606a042ae7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.788541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f23ed48-2000-4480-9c48-0cd09bfaab54 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.820797] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81edcffd-6fc8-4d1b-8366-6227c4fe69c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.831857] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3253462c-b64e-45aa-a74d-c4cf93f47885 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.851097] env[61964]: DEBUG nova.compute.provider_tree [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.356165] env[61964]: DEBUG nova.scheduler.client.report [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 522.705861] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 522.743171] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 522.743461] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.743613] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 522.743788] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.743928] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 522.746309] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 522.746583] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 522.746749] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 522.747175] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 522.747416] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 522.747510] env[61964]: DEBUG nova.virt.hardware [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 522.749162] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cd54b3-f250-4eab-a51c-75d5a745c8dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.758439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0167b16-e1f2-4ace-99b5-f247012e8ac8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.782161] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317dce82-2442-4ec6-88e3-0004838983b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.866467] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.867559] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 522.871104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.560s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.872896] env[61964]: INFO nova.compute.claims [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.379143] env[61964]: DEBUG nova.compute.utils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.391221] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 523.478353] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Successfully created port: 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.790033] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquiring lock "526c58ea-0813-4d49-99cf-11d32bfcca30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.790033] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "526c58ea-0813-4d49-99cf-11d32bfcca30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.892324] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 524.042126] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d30a758-d240-4a51-9967-82819079c39e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.056401] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50010af6-1892-4101-baeb-f046b50c5c13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.097839] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8569f252-c63a-4f67-be21-fb89760362bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.105751] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac5001f-a028-4db7-a6d3-418f73fd682d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.121362] env[61964]: DEBUG nova.compute.provider_tree [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.293593] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 524.629612] env[61964]: DEBUG nova.scheduler.client.report [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 524.825379] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.912695] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 524.958816] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 524.959087] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.959295] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 524.959498] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.959698] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 524.959847] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 524.960112] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 524.960361] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 524.960588] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 524.961037] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 524.961614] env[61964]: DEBUG nova.virt.hardware [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 524.963340] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac9cce7-40fe-4415-a608-1ea6cb1fbbd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.976780] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd60fa73-3e6e-4b27-bc46-7137f1e91546 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.002091] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 525.015269] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.015521] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ecfb715-e834-4e9f-a00a-9b20afe3521d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.029883] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created folder: OpenStack in parent group-v4. [ 525.029883] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating folder: Project (f6a93aba2b6143088dc221bf17e6515b). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.029883] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-987542f3-05b6-4a6b-9eed-8ef9c90a40e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.039924] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created folder: Project (f6a93aba2b6143088dc221bf17e6515b) in parent group-v230360. [ 525.039924] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating folder: Instances. Parent ref: group-v230361. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.039924] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a561f76-f04e-4474-8f66-cfb21169ec30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.049101] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created folder: Instances in parent group-v230361. [ 525.049324] env[61964]: DEBUG oslo.service.loopingcall [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.049526] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 525.049840] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09312213-8113-4532-94b5-00a5d7327358 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.068358] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 525.068358] env[61964]: value = "task-1040499" [ 525.068358] env[61964]: _type = "Task" [ 525.068358] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.077787] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040499, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.135674] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.136242] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 525.141749] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.729s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.143951] env[61964]: INFO nova.compute.claims [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.577805] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040499, 'name': CreateVM_Task, 'duration_secs': 0.313045} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.577805] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 525.578772] env[61964]: DEBUG oslo_vmware.service [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3bc3ef-463f-4b9f-9875-b840133d5470 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.585761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.585761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.585761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 525.585761] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2f598ea-5ea7-49c8-a891-34054928c111 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.591366] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 525.591366] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525301b0-3203-308f-9d7e-ce53457870f0" [ 525.591366] env[61964]: _type = "Task" [ 525.591366] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.598410] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525301b0-3203-308f-9d7e-ce53457870f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.648872] env[61964]: DEBUG nova.compute.utils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.658154] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 525.658154] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 525.706846] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquiring lock "6317e489-9347-46b7-989a-f543af500b39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.707201] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "6317e489-9347-46b7-989a-f543af500b39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.935242] env[61964]: DEBUG nova.policy [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16e252e3ffb846d28fdab47a2e99a0e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38a8b6a394e54168be8155c43a43f35e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.105231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.105510] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 526.105755] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.105885] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.106321] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 526.106582] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d740bcfa-1aed-4609-91ee-ff7fdfd05aaf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.116160] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 526.116349] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 526.117377] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bb8ce9-9ef6-4b22-956f-b0189743377d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.126417] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-588987ef-1418-47b0-90d5-d980dd2ee4cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.134340] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 526.134340] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5283c855-de0c-ae75-f8cd-7c372948e884" [ 526.134340] env[61964]: _type = "Task" [ 526.134340] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.146592] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5283c855-de0c-ae75-f8cd-7c372948e884, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.159319] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 526.211531] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 526.317432] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920bcb78-f8d2-4210-8297-d3c7aa75b909 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.328408] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764c4a64-9074-4dac-adf5-90141046bb2a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.365210] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596d09e0-3cd5-40c7-9ad4-b865c90867d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.373074] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acc1216-369f-4b82-99fd-083b3d9278f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.386857] env[61964]: DEBUG nova.compute.provider_tree [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.644799] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 526.645328] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating directory with path [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 526.645603] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f38c856f-bf46-48e0-a901-15e333930a3e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.673181] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created directory with path [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 526.677123] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Fetch image to [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 526.677123] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloading image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk on the data store datastore2 {{(pid=61964) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 526.677123] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bec8417-7bae-429a-9f7e-191b43f7d1b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.681627] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83e88aa-5815-404c-af72-e07da61d85b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.692683] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43c9910-5001-46fb-b2e1-3c3b6df12262 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.732821] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5efe428-3e1b-4ee1-a307-f31779ac2619 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.742737] env[61964]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-84a9c78a-d7fb-41a1-841b-ab0f9459e4a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.744043] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.771320] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloading image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to the data store datastore2 {{(pid=61964) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 526.855046] env[61964]: DEBUG oslo_vmware.rw_handles [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 526.935953] env[61964]: DEBUG nova.scheduler.client.report [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 527.171907] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 527.207856] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 527.208119] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.208295] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 527.208482] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.208630] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 527.208776] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 527.208982] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 527.210182] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 527.210182] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 527.210314] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 527.210488] env[61964]: DEBUG nova.virt.hardware [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 527.211570] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f2b0eb-2831-4cbe-b757-f2de3c42bfda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.223755] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb1185a-4584-42f0-8b77-f0d18600d997 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.443987] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.443987] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 527.449428] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.624s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.451541] env[61964]: INFO nova.compute.claims [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.570321] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Successfully created port: 326413a6-397d-409b-881a-d0dffec7e92d {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.617316] env[61964]: DEBUG oslo_vmware.rw_handles [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 527.618418] env[61964]: DEBUG oslo_vmware.rw_handles [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 527.683648] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloaded image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk on the data store datastore2 {{(pid=61964) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 527.685533] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 527.685796] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copying Virtual Disk [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk to [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 527.686691] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2fed7f58-572e-4065-b52c-df9d2f33a576 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.695504] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 527.695504] env[61964]: value = "task-1040500" [ 527.695504] env[61964]: _type = "Task" [ 527.695504] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.702862] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.949524] env[61964]: DEBUG nova.compute.utils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.951583] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 527.951793] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.208047] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040500, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.210543] env[61964]: DEBUG nova.policy [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b20910460ad240aeb9b702fd4bd0d160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c117df1b68db404cb7009e76ebeecc93', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.455515] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 528.602381] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b68988e-0eeb-4587-86f6-46e7145c2c84 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.612607] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942fa78b-2d14-4c10-9d90-a79837bc9376 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.651822] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5fa2ad-cb32-4346-8502-a8162917e3f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.664079] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e29f8f9-0d5b-4872-970f-f71abc05ac61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.687892] env[61964]: DEBUG nova.compute.provider_tree [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.710025] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.682264} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.710025] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copied Virtual Disk [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk to [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 528.710025] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 528.710256] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df360177-c46e-4bd5-9b6b-78d1b2ffa519 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.717922] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 528.717922] env[61964]: value = "task-1040501" [ 528.717922] env[61964]: _type = "Task" [ 528.717922] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.726441] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.100617] env[61964]: ERROR nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 529.100617] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.100617] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.100617] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.100617] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.100617] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.100617] env[61964]: ERROR nova.compute.manager raise self.value [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.100617] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.100617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.100617] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.101617] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.101617] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.101617] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 529.101617] env[61964]: ERROR nova.compute.manager [ 529.101617] env[61964]: Traceback (most recent call last): [ 529.101617] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.101617] env[61964]: listener.cb(fileno) [ 529.101617] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.101617] env[61964]: result = function(*args, **kwargs) [ 529.101617] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.101617] env[61964]: return func(*args, **kwargs) [ 529.101617] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.101617] env[61964]: raise e [ 529.101617] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.101617] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 529.101617] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.101617] env[61964]: created_port_ids = self._update_ports_for_instance( [ 529.101617] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.101617] env[61964]: with excutils.save_and_reraise_exception(): [ 529.101617] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.101617] env[61964]: self.force_reraise() [ 529.101617] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.101617] env[61964]: raise self.value [ 529.101617] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.101617] env[61964]: updated_port = self._update_port( [ 529.101617] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.101617] env[61964]: _ensure_no_port_binding_failure(port) [ 529.101617] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.101617] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.103574] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 529.103574] env[61964]: Removing descriptor: 15 [ 529.103574] env[61964]: ERROR nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Traceback (most recent call last): [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] yield resources [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.driver.spawn(context, instance, image_meta, [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.103574] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] vm_ref = self.build_virtual_machine(instance, [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] for vif in network_info: [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self._sync_wrapper(fn, *args, **kwargs) [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.wait() [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self[:] = self._gt.wait() [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self._exit_event.wait() [ 529.104685] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] result = hub.switch() [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self.greenlet.switch() [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] result = function(*args, **kwargs) [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return func(*args, **kwargs) [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise e [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] nwinfo = self.network_api.allocate_for_instance( [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.106571] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] created_port_ids = self._update_ports_for_instance( [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] with excutils.save_and_reraise_exception(): [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.force_reraise() [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise self.value [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] updated_port = self._update_port( [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] _ensure_no_port_binding_failure(port) [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.106933] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise exception.PortBindingFailed(port_id=port['id']) [ 529.107394] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 529.107394] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] [ 529.107394] env[61964]: INFO nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Terminating instance [ 529.126506] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquiring lock "894ad0a3-851a-4b5e-82b0-050cbbf264f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.126769] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "894ad0a3-851a-4b5e-82b0-050cbbf264f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.191932] env[61964]: DEBUG nova.scheduler.client.report [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 529.228641] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022334} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.228913] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 529.229406] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Moving file from [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00/5f28385d-6ea0-420d-8a26-4cb693714c14 to [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14. {{(pid=61964) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 529.229406] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-9893b5b1-cc3f-4ecb-a083-6840012eff51 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.236929] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 529.236929] env[61964]: value = "task-1040502" [ 529.236929] env[61964]: _type = "Task" [ 529.236929] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.246683] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040502, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.469272] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 529.498024] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 529.498024] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.498024] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 529.498230] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.498610] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 529.500369] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 529.500369] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 529.500369] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 529.500369] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 529.500369] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 529.500572] env[61964]: DEBUG nova.virt.hardware [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 529.501431] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2d914e-ad2e-46a4-b841-ccc99ed820ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.515135] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b03188-7c6a-47ce-b81d-549da2ed8544 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.610029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquiring lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.610029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquired lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.610029] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.631230] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 529.699714] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.700316] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 529.704559] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Successfully created port: e7818c75-73c2-4809-9d4b-6c43293258c4 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.706454] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.962s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.707859] env[61964]: INFO nova.compute.claims [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.749766] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040502, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025228} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.749766] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] File moved {{(pid=61964) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 529.749766] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Cleaning up location [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 529.749766] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore2] vmware_temp/64a76d80-af50-4aac-bbe5-e9c6110edd00 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.749766] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee42d177-07f9-48fc-9340-548a6c95475e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.756158] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 529.756158] env[61964]: value = "task-1040503" [ 529.756158] env[61964]: _type = "Task" [ 529.756158] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.765975] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040503, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.996770] env[61964]: ERROR nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 529.996770] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.996770] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.996770] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.996770] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.996770] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.996770] env[61964]: ERROR nova.compute.manager raise self.value [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.996770] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.996770] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.996770] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.997230] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.997230] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.997230] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 529.997230] env[61964]: ERROR nova.compute.manager [ 529.997230] env[61964]: Traceback (most recent call last): [ 529.997230] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.997230] env[61964]: listener.cb(fileno) [ 529.997230] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.997230] env[61964]: result = function(*args, **kwargs) [ 529.997230] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.997230] env[61964]: return func(*args, **kwargs) [ 529.997230] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.997230] env[61964]: raise e [ 529.997230] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.997230] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 529.997230] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.997230] env[61964]: created_port_ids = self._update_ports_for_instance( [ 529.997230] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.997230] env[61964]: with excutils.save_and_reraise_exception(): [ 529.997230] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.997230] env[61964]: self.force_reraise() [ 529.997230] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.997230] env[61964]: raise self.value [ 529.997230] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.997230] env[61964]: updated_port = self._update_port( [ 529.997230] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.997230] env[61964]: _ensure_no_port_binding_failure(port) [ 529.997230] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.997230] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.997933] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 529.997933] env[61964]: Removing descriptor: 17 [ 529.997933] env[61964]: ERROR nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Traceback (most recent call last): [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] yield resources [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.driver.spawn(context, instance, image_meta, [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.997933] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] vm_ref = self.build_virtual_machine(instance, [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] for vif in network_info: [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self._sync_wrapper(fn, *args, **kwargs) [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.wait() [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self[:] = self._gt.wait() [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self._exit_event.wait() [ 529.998261] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] result = hub.switch() [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self.greenlet.switch() [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] result = function(*args, **kwargs) [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return func(*args, **kwargs) [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise e [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] nwinfo = self.network_api.allocate_for_instance( [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.998844] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] created_port_ids = self._update_ports_for_instance( [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] with excutils.save_and_reraise_exception(): [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.force_reraise() [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise self.value [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] updated_port = self._update_port( [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] _ensure_no_port_binding_failure(port) [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.999254] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise exception.PortBindingFailed(port_id=port['id']) [ 529.999566] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 529.999566] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] [ 529.999566] env[61964]: INFO nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Terminating instance [ 530.155764] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.162781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.213729] env[61964]: DEBUG nova.compute.utils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.219064] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 530.221318] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.256020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquiring lock "51b42c8a-f780-4725-85ab-3fad0d621dfd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.256270] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "51b42c8a-f780-4725-85ab-3fad0d621dfd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.272911] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040503, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024159} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.273687] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 530.275266] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0730fcf-ac33-4e05-95b3-bd46029819c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.287294] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 530.287294] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522e05bb-6423-bf94-d45a-4b3e52a5c2c0" [ 530.287294] env[61964]: _type = "Task" [ 530.287294] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.304237] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522e05bb-6423-bf94-d45a-4b3e52a5c2c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010153} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.304836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.305213] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 530.305493] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41e0158d-6a74-4730-af9e-177f78688dc6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.312976] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 530.312976] env[61964]: value = "task-1040504" [ 530.312976] env[61964]: _type = "Task" [ 530.312976] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.321995] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.377172] env[61964]: DEBUG nova.policy [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60f7076ebc0b4f80801c54e91612bc31', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '140ad1faae2d468b8d0a40d1b9e2162e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.504870] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquiring lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.505087] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquired lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.505283] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 530.601248] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.720775] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 530.758853] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 530.779866] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.782683] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.782683] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 530.782683] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 530.831177] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040504, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.892387] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96085d1-50d6-4463-878b-873a0db242e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.900939] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473aeff2-987a-4dc1-9002-296f9d3f5fe5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.937805] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c4d537-63af-420a-bb98-4167f283a689 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.946136] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bb0090-2ffd-465f-80e3-9737a18fd0b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.960705] env[61964]: DEBUG nova.compute.provider_tree [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.049320] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.108502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Releasing lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.108935] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 531.109159] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 531.109472] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-934f13a0-1efb-4b2d-80fd-c90338134255 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.121298] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262670f1-e968-4014-acb8-c6f9d26e8ad9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.144836] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e23a7d59-0516-46b9-ab6d-e4c313ad758a could not be found. [ 531.145085] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 531.145769] env[61964]: INFO nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 531.146070] env[61964]: DEBUG oslo.service.loopingcall [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.146298] env[61964]: DEBUG nova.compute.manager [-] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 531.146383] env[61964]: DEBUG nova.network.neutron [-] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 531.196335] env[61964]: DEBUG nova.network.neutron [-] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.208469] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Successfully created port: ab3a7988-20c1-486c-8b69-9c47f0780662 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.218485] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.286442] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.286618] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.286743] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.286864] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.286980] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.287108] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6317e489-9347-46b7-989a-f543af500b39] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 531.287227] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Didn't find any instances for network info cache update. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10456}} [ 531.287693] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.287934] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.288448] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.288671] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.288873] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.289080] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.289245] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 531.289390] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.292148] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.327361] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548294} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.327585] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 531.327791] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 531.328095] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdab12fe-fde5-414c-b51d-1c4468a94a52 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.334155] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 531.334155] env[61964]: value = "task-1040505" [ 531.334155] env[61964]: _type = "Task" [ 531.334155] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.343148] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.376101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "ee058557-cace-4e32-bdf9-bd4af7fe29ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.376620] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "ee058557-cace-4e32-bdf9-bd4af7fe29ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.466033] env[61964]: DEBUG nova.scheduler.client.report [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 531.701688] env[61964]: DEBUG nova.network.neutron [-] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.725017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Releasing lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.725017] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 531.725017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 531.725017] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6b1acc3-83de-4b8c-afa8-37418f639605 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.733636] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c960bc87-a4e8-4c0c-82aa-60031f16669d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.749868] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 531.772019] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5ee1938-27ad-4189-b9f0-9c9821e0ffc8 could not be found. [ 531.772019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 531.772019] env[61964]: INFO nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Took 0.05 seconds to destroy the instance on the hypervisor. [ 531.772019] env[61964]: DEBUG oslo.service.loopingcall [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.772019] env[61964]: DEBUG nova.compute.manager [-] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 531.772019] env[61964]: DEBUG nova.network.neutron [-] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 531.784947] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 531.785200] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.785363] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 531.785538] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.785687] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 531.785832] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 531.786099] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 531.786507] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 531.786706] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 531.786871] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 531.787053] env[61964]: DEBUG nova.virt.hardware [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 531.788245] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79444cd6-98dc-462b-a245-e32b1233d5eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.792553] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.796534] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261f6b20-0173-4b52-84fc-5f8f03b884c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.812480] env[61964]: DEBUG nova.network.neutron [-] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.844547] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109103} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.844779] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 531.845717] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136d4e3b-c60a-46ca-a61c-1b3b92746b1c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.869341] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 531.869524] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee3ca2bc-3b5d-4250-ba18-490c0dbd5989 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.885260] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 531.894461] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 531.894461] env[61964]: value = "task-1040506" [ 531.894461] env[61964]: _type = "Task" [ 531.894461] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.903250] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.968512] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.969060] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 531.972330] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.810s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.974120] env[61964]: INFO nova.compute.claims [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.205421] env[61964]: INFO nova.compute.manager [-] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Took 1.06 seconds to deallocate network for instance. [ 532.207785] env[61964]: DEBUG nova.compute.claims [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.208007] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.281709] env[61964]: DEBUG nova.compute.manager [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Received event network-changed-765a8f0b-e2e8-4ce9-84fd-c54a12cc0465 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 532.281709] env[61964]: DEBUG nova.compute.manager [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Refreshing instance network info cache due to event network-changed-765a8f0b-e2e8-4ce9-84fd-c54a12cc0465. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 532.281709] env[61964]: DEBUG oslo_concurrency.lockutils [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] Acquiring lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.281709] env[61964]: DEBUG oslo_concurrency.lockutils [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] Acquired lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.281709] env[61964]: DEBUG nova.network.neutron [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Refreshing network info cache for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 532.316198] env[61964]: DEBUG nova.network.neutron [-] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.410726] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040506, 'name': ReconfigVM_Task, 'duration_secs': 0.278448} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.411079] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 532.411826] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bcc54ed9-5d74-4fb1-b5e1-1513d348a348 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.416515] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.421837] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 532.421837] env[61964]: value = "task-1040507" [ 532.421837] env[61964]: _type = "Task" [ 532.421837] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.431089] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040507, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.473933] env[61964]: DEBUG nova.compute.utils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.475459] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 532.475595] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.580164] env[61964]: DEBUG nova.policy [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6f362bd9f5840ba894ea5787ee257b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68ead29946cd49689a2bf3c87df677eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.820140] env[61964]: INFO nova.compute.manager [-] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Took 1.05 seconds to deallocate network for instance. [ 532.823942] env[61964]: DEBUG nova.compute.claims [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.823942] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.834467] env[61964]: DEBUG nova.network.neutron [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.920163] env[61964]: ERROR nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 532.920163] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.920163] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.920163] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.920163] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.920163] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.920163] env[61964]: ERROR nova.compute.manager raise self.value [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.920163] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.920163] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.920163] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.920858] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.920858] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.920858] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 532.920858] env[61964]: ERROR nova.compute.manager [ 532.920858] env[61964]: Traceback (most recent call last): [ 532.920858] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.920858] env[61964]: listener.cb(fileno) [ 532.920858] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.920858] env[61964]: result = function(*args, **kwargs) [ 532.920858] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.920858] env[61964]: return func(*args, **kwargs) [ 532.920858] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 532.920858] env[61964]: raise e [ 532.920858] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.920858] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 532.920858] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.920858] env[61964]: created_port_ids = self._update_ports_for_instance( [ 532.920858] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.920858] env[61964]: with excutils.save_and_reraise_exception(): [ 532.920858] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.920858] env[61964]: self.force_reraise() [ 532.920858] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.920858] env[61964]: raise self.value [ 532.920858] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.920858] env[61964]: updated_port = self._update_port( [ 532.920858] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.920858] env[61964]: _ensure_no_port_binding_failure(port) [ 532.920858] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.920858] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.922187] env[61964]: nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 532.922187] env[61964]: Removing descriptor: 16 [ 532.922187] env[61964]: ERROR nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Traceback (most recent call last): [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] yield resources [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.driver.spawn(context, instance, image_meta, [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.922187] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] vm_ref = self.build_virtual_machine(instance, [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] for vif in network_info: [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self._sync_wrapper(fn, *args, **kwargs) [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.wait() [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self[:] = self._gt.wait() [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self._exit_event.wait() [ 532.922733] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] result = hub.switch() [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self.greenlet.switch() [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] result = function(*args, **kwargs) [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return func(*args, **kwargs) [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise e [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] nwinfo = self.network_api.allocate_for_instance( [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.925487] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] created_port_ids = self._update_ports_for_instance( [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] with excutils.save_and_reraise_exception(): [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.force_reraise() [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise self.value [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] updated_port = self._update_port( [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] _ensure_no_port_binding_failure(port) [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.926421] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise exception.PortBindingFailed(port_id=port['id']) [ 532.927366] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 532.927366] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] [ 532.927366] env[61964]: INFO nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Terminating instance [ 532.946941] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040507, 'name': Rename_Task, 'duration_secs': 0.138913} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.948335] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 532.948464] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-385c6857-3111-4f44-8ddf-9c028756bd37 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.957435] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 532.957435] env[61964]: value = "task-1040508" [ 532.957435] env[61964]: _type = "Task" [ 532.957435] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.970679] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040508, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.972245] env[61964]: ERROR nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 532.972245] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.972245] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.972245] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.972245] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.972245] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.972245] env[61964]: ERROR nova.compute.manager raise self.value [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.972245] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.972245] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.972245] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.972843] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.972843] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.972843] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 532.972843] env[61964]: ERROR nova.compute.manager [ 532.972843] env[61964]: Traceback (most recent call last): [ 532.972843] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.972843] env[61964]: listener.cb(fileno) [ 532.972843] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.972843] env[61964]: result = function(*args, **kwargs) [ 532.972843] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.972843] env[61964]: return func(*args, **kwargs) [ 532.972843] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 532.972843] env[61964]: raise e [ 532.972843] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.972843] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 532.972843] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.972843] env[61964]: created_port_ids = self._update_ports_for_instance( [ 532.972843] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.972843] env[61964]: with excutils.save_and_reraise_exception(): [ 532.972843] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.972843] env[61964]: self.force_reraise() [ 532.972843] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.972843] env[61964]: raise self.value [ 532.972843] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.972843] env[61964]: updated_port = self._update_port( [ 532.972843] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.972843] env[61964]: _ensure_no_port_binding_failure(port) [ 532.972843] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.972843] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.973610] env[61964]: nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 532.973610] env[61964]: Removing descriptor: 15 [ 532.973610] env[61964]: ERROR nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Traceback (most recent call last): [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] yield resources [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.driver.spawn(context, instance, image_meta, [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.973610] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] vm_ref = self.build_virtual_machine(instance, [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] for vif in network_info: [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self._sync_wrapper(fn, *args, **kwargs) [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.wait() [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self[:] = self._gt.wait() [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self._exit_event.wait() [ 532.973920] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] result = hub.switch() [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self.greenlet.switch() [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] result = function(*args, **kwargs) [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return func(*args, **kwargs) [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise e [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] nwinfo = self.network_api.allocate_for_instance( [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.974279] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] created_port_ids = self._update_ports_for_instance( [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] with excutils.save_and_reraise_exception(): [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.force_reraise() [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise self.value [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] updated_port = self._update_port( [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] _ensure_no_port_binding_failure(port) [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.974587] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise exception.PortBindingFailed(port_id=port['id']) [ 532.974871] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 532.974871] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] [ 532.974871] env[61964]: INFO nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Terminating instance [ 532.986215] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 533.115926] env[61964]: DEBUG nova.network.neutron [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.169892] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Successfully created port: c42144e7-4206-4859-8d44-6f52bb282ffe {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.256342] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5883e1-2cea-46f9-b3d6-5b5ec4aa3147 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.268022] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc9c84e-942a-42aa-afff-a242e0770bf2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.306587] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451e2371-0394-4935-9425-05eb2f584c24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.313539] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c1125a-43e4-4b95-b344-032f37a581e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.334464] env[61964]: DEBUG nova.compute.provider_tree [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.441031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.441031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquired lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.441031] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.471599] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040508, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.490281] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquiring lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.490567] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquired lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.490813] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.619522] env[61964]: DEBUG oslo_concurrency.lockutils [req-d1b000c5-3879-4f13-8309-9a0fd5a9f1f6 req-22ca9b72-16b9-4c5f-820d-225033072445 service nova] Releasing lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.648823] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquiring lock "c9e79aac-19af-4339-9b06-5907d2ac716c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.650792] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "c9e79aac-19af-4339-9b06-5907d2ac716c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.838105] env[61964]: DEBUG nova.scheduler.client.report [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 533.969451] env[61964]: DEBUG oslo_vmware.api [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040508, 'name': PowerOnVM_Task, 'duration_secs': 0.837995} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.969904] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 533.970217] env[61964]: INFO nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Took 9.06 seconds to spawn the instance on the hypervisor. [ 533.970675] env[61964]: DEBUG nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 533.971568] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95951e21-7036-4dd0-8875-0084f839a89e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.998027] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 534.039480] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.048208] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 534.048208] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.048435] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 534.049913] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.049913] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 534.049913] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 534.049913] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 534.049913] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 534.050135] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 534.050135] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 534.050135] env[61964]: DEBUG nova.virt.hardware [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 534.051211] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b9c3f7-4adf-41d1-96d6-d2950e690e07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.064628] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d739e7-3720-4d1e-90f0-c9ab37205237 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.073540] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.121185] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.154134] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 534.273202] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.347794] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.349076] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 534.355327] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.062s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.357807] env[61964]: INFO nova.compute.claims [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.493470] env[61964]: INFO nova.compute.manager [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Took 14.38 seconds to build instance. [ 534.538743] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "07771b38-3dc9-4608-8bea-48971bbc9651" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.539172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "07771b38-3dc9-4608-8bea-48971bbc9651" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.623259] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Releasing lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.623714] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 534.623935] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 534.624681] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbf94587-8bfc-4edb-a9ed-7e71aa0cdd42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.634071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fd9e28-11ce-4f20-a582-4cb672e4ff51 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.657338] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ec4c677-9166-4b3c-ade7-548bee8e245b could not be found. [ 534.657580] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.657767] env[61964]: INFO nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 534.658025] env[61964]: DEBUG oslo.service.loopingcall [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.658242] env[61964]: DEBUG nova.compute.manager [-] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 534.658352] env[61964]: DEBUG nova.network.neutron [-] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.678811] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.689026] env[61964]: DEBUG nova.network.neutron [-] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.775768] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Releasing lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.775768] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 534.776071] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 534.776358] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a45fa84-64c4-462a-90cb-02e70d74e03e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.786797] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f732ad25-6036-466d-a7e8-07838f68f605 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.814919] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 526c58ea-0813-4d49-99cf-11d32bfcca30 could not be found. [ 534.815910] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.815910] env[61964]: INFO nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Took 0.04 seconds to destroy the instance on the hypervisor. [ 534.816041] env[61964]: DEBUG oslo.service.loopingcall [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.816480] env[61964]: DEBUG nova.compute.manager [-] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 534.816605] env[61964]: DEBUG nova.network.neutron [-] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.840767] env[61964]: DEBUG nova.network.neutron [-] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.867513] env[61964]: DEBUG nova.compute.utils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.867513] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 534.867513] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 534.995293] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a3aa313f-a511-41a5-a12e-57d569c68be8 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.916s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.013165] env[61964]: DEBUG nova.policy [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20ae2c0542f740bc950c49a5be8ee135', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '990010584ccc4fd496c9ca45411f8171', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.055965] env[61964]: DEBUG nova.compute.manager [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Received event network-changed-e7818c75-73c2-4809-9d4b-6c43293258c4 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 535.056879] env[61964]: DEBUG nova.compute.manager [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Refreshing instance network info cache due to event network-changed-e7818c75-73c2-4809-9d4b-6c43293258c4. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 535.057176] env[61964]: DEBUG oslo_concurrency.lockutils [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] Acquiring lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.057287] env[61964]: DEBUG oslo_concurrency.lockutils [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] Acquired lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.057430] env[61964]: DEBUG nova.network.neutron [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Refreshing network info cache for port e7818c75-73c2-4809-9d4b-6c43293258c4 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.192309] env[61964]: DEBUG nova.network.neutron [-] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.346973] env[61964]: DEBUG nova.network.neutron [-] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.373097] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 535.501815] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 535.563613] env[61964]: ERROR nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 535.563613] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 535.563613] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.563613] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.563613] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.563613] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.563613] env[61964]: ERROR nova.compute.manager raise self.value [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.563613] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.563613] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.563613] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.564298] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.564298] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.564298] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 535.564298] env[61964]: ERROR nova.compute.manager [ 535.564298] env[61964]: Traceback (most recent call last): [ 535.564298] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.564298] env[61964]: listener.cb(fileno) [ 535.564298] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.564298] env[61964]: result = function(*args, **kwargs) [ 535.564298] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.564298] env[61964]: return func(*args, **kwargs) [ 535.564298] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 535.564298] env[61964]: raise e [ 535.564298] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 535.564298] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 535.564298] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.564298] env[61964]: created_port_ids = self._update_ports_for_instance( [ 535.564298] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.564298] env[61964]: with excutils.save_and_reraise_exception(): [ 535.564298] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.564298] env[61964]: self.force_reraise() [ 535.564298] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.564298] env[61964]: raise self.value [ 535.564298] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.564298] env[61964]: updated_port = self._update_port( [ 535.564298] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.564298] env[61964]: _ensure_no_port_binding_failure(port) [ 535.564298] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.564298] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.564983] env[61964]: nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 535.564983] env[61964]: Removing descriptor: 17 [ 535.566168] env[61964]: ERROR nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] Traceback (most recent call last): [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] yield resources [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.driver.spawn(context, instance, image_meta, [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] vm_ref = self.build_virtual_machine(instance, [ 535.566168] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] for vif in network_info: [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self._sync_wrapper(fn, *args, **kwargs) [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.wait() [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self[:] = self._gt.wait() [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self._exit_event.wait() [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.566839] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] result = hub.switch() [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self.greenlet.switch() [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] result = function(*args, **kwargs) [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return func(*args, **kwargs) [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise e [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] nwinfo = self.network_api.allocate_for_instance( [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] created_port_ids = self._update_ports_for_instance( [ 535.567158] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] with excutils.save_and_reraise_exception(): [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.force_reraise() [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise self.value [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] updated_port = self._update_port( [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] _ensure_no_port_binding_failure(port) [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise exception.PortBindingFailed(port_id=port['id']) [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 535.567528] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] [ 535.567877] env[61964]: INFO nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Terminating instance [ 535.614217] env[61964]: DEBUG nova.network.neutron [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.621970] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4090e8ea-c654-4e97-9447-10bee9715698 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.630839] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b85bf2-9422-4ca4-8e10-8f6c4137fc00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.678271] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914db04f-c087-4746-b53e-9620cd52cb9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.684971] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquiring lock "bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.685479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.690738] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Successfully created port: 98482e59-e602-4e29-9a8a-ab8c13d042a6 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.693658] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac5eefa-52e0-4a00-8497-99fe0d42eeb5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.698716] env[61964]: INFO nova.compute.manager [-] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Took 1.04 seconds to deallocate network for instance. [ 535.701009] env[61964]: DEBUG nova.compute.claims [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.701180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.710517] env[61964]: DEBUG nova.compute.provider_tree [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.761343] env[61964]: DEBUG nova.network.neutron [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.850792] env[61964]: INFO nova.compute.manager [-] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Took 1.03 seconds to deallocate network for instance. [ 535.854115] env[61964]: DEBUG nova.compute.claims [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.854241] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.991597] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Received event network-vif-deleted-765a8f0b-e2e8-4ce9-84fd-c54a12cc0465 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 535.991851] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Received event network-changed-326413a6-397d-409b-881a-d0dffec7e92d {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 535.991981] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Refreshing instance network info cache due to event network-changed-326413a6-397d-409b-881a-d0dffec7e92d. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 535.992682] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Acquiring lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.992848] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Acquired lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.993032] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Refreshing network info cache for port 326413a6-397d-409b-881a-d0dffec7e92d {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 536.031035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.074237] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquiring lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.074237] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquired lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.074237] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.215590] env[61964]: DEBUG nova.scheduler.client.report [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 536.263566] env[61964]: DEBUG oslo_concurrency.lockutils [req-fd93844d-1d29-43d6-8c41-82b1bd43be82 req-0e33c493-87d1-4a6e-840d-305de4d8c66b service nova] Releasing lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.394540] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 536.440298] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 536.440298] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.440298] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 536.440298] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.440545] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 536.440545] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 536.440545] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 536.440545] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 536.440545] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 536.440686] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 536.440686] env[61964]: DEBUG nova.virt.hardware [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 536.441282] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c3598c-1dea-414a-be9b-3ca67f89772a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.450093] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a975a4a-c300-4841-9413-bc3ca74a2dd4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.528864] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.606460] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.606684] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.608521] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.683276] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.722426] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.722426] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 536.724361] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.932s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.724539] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.724680] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 536.724977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.517s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.731018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9879168c-ef73-4060-8bb2-96d9ff64d936 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.748611] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b9560e-c8f1-42ae-90ab-ad673d0d9716 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.767869] env[61964]: INFO nova.compute.manager [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Rebuilding instance [ 536.770894] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.772969] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96301046-b833-47a0-9533-bfa39421a220 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.780605] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399e944f-99a6-4090-81c3-6611d5907a6b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.816367] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181613MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 536.816532] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.902292] env[61964]: DEBUG nova.compute.manager [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 536.903858] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751eec4e-c65e-43e4-beb9-4ce3d217afff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.189121] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Releasing lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.189428] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Received event network-vif-deleted-326413a6-397d-409b-881a-d0dffec7e92d {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 537.189608] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Received event network-changed-ab3a7988-20c1-486c-8b69-9c47f0780662 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 537.189765] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Refreshing instance network info cache due to event network-changed-ab3a7988-20c1-486c-8b69-9c47f0780662. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 537.189971] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Acquiring lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.190123] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Acquired lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.190280] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Refreshing network info cache for port ab3a7988-20c1-486c-8b69-9c47f0780662 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 537.243344] env[61964]: DEBUG nova.compute.utils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.243344] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 537.243344] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.276605] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Releasing lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.277051] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 537.277247] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 537.277529] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fe7c67a-6211-435c-8660-8678603db066 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.291381] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f44fecb-662a-4a9e-98d2-fc7f5aada401 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.325167] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6317e489-9347-46b7-989a-f543af500b39 could not be found. [ 537.325167] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 537.325464] env[61964]: INFO nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Took 0.05 seconds to destroy the instance on the hypervisor. [ 537.325715] env[61964]: DEBUG oslo.service.loopingcall [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.326039] env[61964]: DEBUG nova.compute.manager [-] [instance: 6317e489-9347-46b7-989a-f543af500b39] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 537.326398] env[61964]: DEBUG nova.network.neutron [-] [instance: 6317e489-9347-46b7-989a-f543af500b39] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.348211] env[61964]: DEBUG nova.policy [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '02dad13593ac457a97516748e69a6185', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c50dadca12fb45e9a432bc75724c702b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 537.350716] env[61964]: DEBUG nova.network.neutron [-] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.517783] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e92bbb-a842-48c2-ae2c-bacee1c4cc92 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.525540] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c807b1-a735-4425-a451-138ca9c9df41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.559264] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f95788-498e-49f7-b433-0d7b0f42cb01 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.567479] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5d86a9-b08a-46ee-a7dc-3123a46e9965 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.581143] env[61964]: DEBUG nova.compute.provider_tree [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.715553] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.750723] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 537.769608] env[61964]: ERROR nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 537.769608] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.769608] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.769608] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.769608] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.769608] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.769608] env[61964]: ERROR nova.compute.manager raise self.value [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.769608] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.769608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.769608] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.770072] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.770072] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.770072] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 537.770072] env[61964]: ERROR nova.compute.manager [ 537.770072] env[61964]: Traceback (most recent call last): [ 537.770072] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.770072] env[61964]: listener.cb(fileno) [ 537.770072] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.770072] env[61964]: result = function(*args, **kwargs) [ 537.770072] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.770072] env[61964]: return func(*args, **kwargs) [ 537.770072] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.770072] env[61964]: raise e [ 537.770072] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.770072] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 537.770072] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.770072] env[61964]: created_port_ids = self._update_ports_for_instance( [ 537.770072] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.770072] env[61964]: with excutils.save_and_reraise_exception(): [ 537.770072] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.770072] env[61964]: self.force_reraise() [ 537.770072] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.770072] env[61964]: raise self.value [ 537.770072] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.770072] env[61964]: updated_port = self._update_port( [ 537.770072] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.770072] env[61964]: _ensure_no_port_binding_failure(port) [ 537.770072] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.770072] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.771094] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 537.771094] env[61964]: Removing descriptor: 15 [ 537.771094] env[61964]: ERROR nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Traceback (most recent call last): [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] yield resources [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.driver.spawn(context, instance, image_meta, [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.771094] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] vm_ref = self.build_virtual_machine(instance, [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] for vif in network_info: [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self._sync_wrapper(fn, *args, **kwargs) [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.wait() [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self[:] = self._gt.wait() [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self._exit_event.wait() [ 537.771524] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] result = hub.switch() [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self.greenlet.switch() [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] result = function(*args, **kwargs) [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return func(*args, **kwargs) [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise e [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] nwinfo = self.network_api.allocate_for_instance( [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.771875] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] created_port_ids = self._update_ports_for_instance( [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] with excutils.save_and_reraise_exception(): [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.force_reraise() [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise self.value [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] updated_port = self._update_port( [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] _ensure_no_port_binding_failure(port) [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.772280] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise exception.PortBindingFailed(port_id=port['id']) [ 537.772694] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 537.772694] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] [ 537.772694] env[61964]: INFO nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Terminating instance [ 537.845928] env[61964]: DEBUG nova.network.neutron [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.854679] env[61964]: DEBUG nova.network.neutron [-] [instance: 6317e489-9347-46b7-989a-f543af500b39] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.922674] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 537.923034] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df9559c9-f34e-49c2-821a-a2b86485dcfc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.931807] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 537.931807] env[61964]: value = "task-1040509" [ 537.931807] env[61964]: _type = "Task" [ 537.931807] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.943817] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040509, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.085563] env[61964]: DEBUG nova.scheduler.client.report [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 538.157599] env[61964]: DEBUG nova.compute.manager [req-eb510351-8308-4011-9aa4-849c6a141a64 req-db3edaf1-a465-4654-8448-cbe8f5d756fa service nova] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Received event network-vif-deleted-e7818c75-73c2-4809-9d4b-6c43293258c4 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 538.279729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquiring lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.280503] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquired lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.280503] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.349256] env[61964]: DEBUG oslo_concurrency.lockutils [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] Releasing lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.349256] env[61964]: DEBUG nova.compute.manager [req-b98e9be3-d661-486c-b3d6-0fd5d62c1853 req-96da30fd-4b52-4318-bc80-28a42555f115 service nova] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Received event network-vif-deleted-ab3a7988-20c1-486c-8b69-9c47f0780662 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 538.360188] env[61964]: INFO nova.compute.manager [-] [instance: 6317e489-9347-46b7-989a-f543af500b39] Took 1.03 seconds to deallocate network for instance. [ 538.362763] env[61964]: DEBUG nova.compute.claims [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 538.363016] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.446376] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040509, 'name': PowerOffVM_Task, 'duration_secs': 0.135832} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.446940] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 538.447278] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.449100] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c72e72e-869d-4888-9f79-d2442319b17b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.459132] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 538.459654] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-713f708f-6352-4cbd-81f5-fcf047047a88 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.487296] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 538.487559] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 538.487697] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore2] 4e14a519-b543-4a18-9e1d-7df16abf5345 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 538.487957] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-097df4ed-d1ed-4ba4-9f2d-8cc40ea2c189 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.497612] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 538.497612] env[61964]: value = "task-1040511" [ 538.497612] env[61964]: _type = "Task" [ 538.497612] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.509510] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.545609] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Successfully created port: f4628cb6-2328-403d-aa29-6dc2c4191ef1 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.594758] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.595431] env[61964]: ERROR nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Traceback (most recent call last): [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.driver.spawn(context, instance, image_meta, [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] vm_ref = self.build_virtual_machine(instance, [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.595431] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] for vif in network_info: [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self._sync_wrapper(fn, *args, **kwargs) [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.wait() [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self[:] = self._gt.wait() [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self._exit_event.wait() [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] result = hub.switch() [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.595786] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return self.greenlet.switch() [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] result = function(*args, **kwargs) [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] return func(*args, **kwargs) [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise e [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] nwinfo = self.network_api.allocate_for_instance( [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] created_port_ids = self._update_ports_for_instance( [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] with excutils.save_and_reraise_exception(): [ 538.596115] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] self.force_reraise() [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise self.value [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] updated_port = self._update_port( [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] _ensure_no_port_binding_failure(port) [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] raise exception.PortBindingFailed(port_id=port['id']) [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] nova.exception.PortBindingFailed: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. [ 538.596506] env[61964]: ERROR nova.compute.manager [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] [ 538.597338] env[61964]: DEBUG nova.compute.utils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 538.598649] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.182s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.600259] env[61964]: INFO nova.compute.claims [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.607767] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Build of instance e23a7d59-0516-46b9-ab6d-e4c313ad758a was re-scheduled: Binding failed for port 765a8f0b-e2e8-4ce9-84fd-c54a12cc0465, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 538.608696] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 538.608794] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquiring lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.608937] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Acquired lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.609502] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.762323] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 538.793993] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 538.794312] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 538.795272] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 538.795512] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 538.795765] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 538.796048] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 538.796206] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 538.796367] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 538.796532] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 538.796706] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 538.796881] env[61964]: DEBUG nova.virt.hardware [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 538.797802] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292856e8-6fb2-4299-84a0-82a2ba883435 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.806601] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.809254] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69a40c9-919e-47fa-a376-478e60b1867f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.890519] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.955934] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Received event network-changed-c42144e7-4206-4859-8d44-6f52bb282ffe {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 538.956156] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Refreshing instance network info cache due to event network-changed-c42144e7-4206-4859-8d44-6f52bb282ffe. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 538.956372] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Acquiring lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.956514] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Acquired lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.956676] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Refreshing network info cache for port c42144e7-4206-4859-8d44-6f52bb282ffe {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.007470] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092176} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.007719] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 539.007974] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 539.008177] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 539.133060] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.272378] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.394407] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Releasing lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.395196] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 539.395441] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 539.395752] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dc5f79d-bc36-4508-8dd9-e39b8e18d103 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.405741] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a7a3c9-f821-489b-a4ce-1178102cf66e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.432497] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 894ad0a3-851a-4b5e-82b0-050cbbf264f6 could not be found. [ 539.432497] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 539.432874] env[61964]: INFO nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 539.432874] env[61964]: DEBUG oslo.service.loopingcall [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.433059] env[61964]: DEBUG nova.compute.manager [-] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 539.433059] env[61964]: DEBUG nova.network.neutron [-] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.458473] env[61964]: DEBUG nova.network.neutron [-] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.481851] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.619591] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.781658] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Releasing lock "refresh_cache-e23a7d59-0516-46b9-ab6d-e4c313ad758a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.781946] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 539.782144] env[61964]: DEBUG nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 539.782313] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.805086] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.853096] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1402775-9eab-4788-9b60-ed643f3bf8d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.862959] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce07fe3-d4d5-478d-af0c-b4d40a9ef368 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.897186] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2839dda0-443a-4436-b68f-cdeadcb58803 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.909265] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2ddc14-fcf3-496f-940a-046cbf11ea55 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.923870] env[61964]: DEBUG nova.compute.provider_tree [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.963571] env[61964]: DEBUG nova.network.neutron [-] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.049784] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 540.049784] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.049784] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 540.049784] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.050081] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 540.050081] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 540.050081] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 540.050210] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 540.050372] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 540.050529] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 540.050698] env[61964]: DEBUG nova.virt.hardware [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 540.052824] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26bb588-3ff0-42dc-92d5-de78111cdb71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.060865] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac173c1-43bb-4e42-8a92-546974fb0a17 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.076378] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 540.082176] env[61964]: DEBUG oslo.service.loopingcall [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.083155] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 540.083155] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d7f0505-ea2b-4ed1-a8d3-7d2bdcc2a297 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.103067] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 540.103067] env[61964]: value = "task-1040512" [ 540.103067] env[61964]: _type = "Task" [ 540.103067] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.113962] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040512, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.126707] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Releasing lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.126707] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 6317e489-9347-46b7-989a-f543af500b39] Received event network-vif-deleted-c42144e7-4206-4859-8d44-6f52bb282ffe {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 540.126707] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Received event network-changed-98482e59-e602-4e29-9a8a-ab8c13d042a6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 540.126707] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Refreshing instance network info cache due to event network-changed-98482e59-e602-4e29-9a8a-ab8c13d042a6. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 540.126707] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Acquiring lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.126951] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Acquired lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.126951] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Refreshing network info cache for port 98482e59-e602-4e29-9a8a-ab8c13d042a6 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 540.311140] env[61964]: DEBUG nova.network.neutron [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.427032] env[61964]: DEBUG nova.scheduler.client.report [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 540.468288] env[61964]: INFO nova.compute.manager [-] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Took 1.04 seconds to deallocate network for instance. [ 540.474417] env[61964]: DEBUG nova.compute.claims [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.474502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.616273] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040512, 'name': CreateVM_Task, 'duration_secs': 0.245585} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.616273] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 540.617180] env[61964]: DEBUG oslo_vmware.service [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a384aa-fc37-4981-818a-6038a611e6e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.627008] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.627008] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.627008] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 540.627008] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d89e0815-a2ad-43ca-974d-59f9b89df4c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.632982] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 540.632982] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c72ced-f4f2-1a25-19bf-85ee42559ef5" [ 540.632982] env[61964]: _type = "Task" [ 540.632982] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.643539] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c72ced-f4f2-1a25-19bf-85ee42559ef5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.653465] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.729828] env[61964]: DEBUG nova.network.neutron [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.815018] env[61964]: ERROR nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 540.815018] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.815018] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.815018] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.815018] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.815018] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.815018] env[61964]: ERROR nova.compute.manager raise self.value [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.815018] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.815018] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.815018] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.815605] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.815605] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.815605] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 540.815605] env[61964]: ERROR nova.compute.manager [ 540.815605] env[61964]: Traceback (most recent call last): [ 540.815605] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.815605] env[61964]: listener.cb(fileno) [ 540.815605] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.815605] env[61964]: result = function(*args, **kwargs) [ 540.815605] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.815605] env[61964]: return func(*args, **kwargs) [ 540.815605] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.815605] env[61964]: raise e [ 540.815605] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.815605] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 540.815605] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.815605] env[61964]: created_port_ids = self._update_ports_for_instance( [ 540.815605] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.815605] env[61964]: with excutils.save_and_reraise_exception(): [ 540.815605] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.815605] env[61964]: self.force_reraise() [ 540.815605] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.815605] env[61964]: raise self.value [ 540.815605] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.815605] env[61964]: updated_port = self._update_port( [ 540.815605] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.815605] env[61964]: _ensure_no_port_binding_failure(port) [ 540.815605] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.815605] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.816387] env[61964]: nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 540.816387] env[61964]: Removing descriptor: 17 [ 540.816387] env[61964]: ERROR nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Traceback (most recent call last): [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] yield resources [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.driver.spawn(context, instance, image_meta, [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.816387] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] vm_ref = self.build_virtual_machine(instance, [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] for vif in network_info: [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self._sync_wrapper(fn, *args, **kwargs) [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.wait() [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self[:] = self._gt.wait() [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self._exit_event.wait() [ 540.816724] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] result = hub.switch() [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self.greenlet.switch() [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] result = function(*args, **kwargs) [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return func(*args, **kwargs) [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise e [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] nwinfo = self.network_api.allocate_for_instance( [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.817043] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] created_port_ids = self._update_ports_for_instance( [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] with excutils.save_and_reraise_exception(): [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.force_reraise() [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise self.value [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] updated_port = self._update_port( [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] _ensure_no_port_binding_failure(port) [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.817389] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise exception.PortBindingFailed(port_id=port['id']) [ 540.817730] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 540.817730] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] [ 540.817730] env[61964]: INFO nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Terminating instance [ 540.822063] env[61964]: INFO nova.compute.manager [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] [instance: e23a7d59-0516-46b9-ab6d-e4c313ad758a] Took 1.04 seconds to deallocate network for instance. [ 540.931647] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.932215] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 540.934728] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.111s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.076589] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquiring lock "84cda217-8acd-4fd9-80ac-cb83edcc1347" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.076934] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "84cda217-8acd-4fd9-80ac-cb83edcc1347" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.148418] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.151427] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 541.151427] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.151427] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.151427] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.151427] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebfd6774-7078-49ae-aab6-913fd8dd947d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.167639] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.167906] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 541.168953] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42228e0-9acd-462d-923d-18cec69395bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.176521] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4b63087-bafa-43cf-a795-09e9d0bdbd8b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.181715] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 541.181715] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b61cc3-3645-adc8-e17a-2d927f97ddce" [ 541.181715] env[61964]: _type = "Task" [ 541.181715] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.193490] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b61cc3-3645-adc8-e17a-2d927f97ddce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.233310] env[61964]: DEBUG oslo_concurrency.lockutils [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] Releasing lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.233608] env[61964]: DEBUG nova.compute.manager [req-5cabc6c0-8adc-4630-b605-62012d83825b req-a073bfc6-f4a2-41df-8283-9a517ddf617c service nova] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Received event network-vif-deleted-98482e59-e602-4e29-9a8a-ab8c13d042a6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 541.326163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquiring lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.326163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquired lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.326163] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.445362] env[61964]: DEBUG nova.compute.utils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.448749] env[61964]: DEBUG nova.compute.manager [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Received event network-changed-f4628cb6-2328-403d-aa29-6dc2c4191ef1 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 541.449073] env[61964]: DEBUG nova.compute.manager [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Refreshing instance network info cache due to event network-changed-f4628cb6-2328-403d-aa29-6dc2c4191ef1. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 541.449358] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] Acquiring lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.451420] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 541.454514] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 541.576441] env[61964]: DEBUG nova.policy [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb7cc75be9a44e2187090e490b8056e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833e91665288466a9ac2141d516ab922', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.692582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 541.696311] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating directory with path [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 541.696595] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a8f5839-612f-4096-bd03-bb4bf3140247 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.699988] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebd7f50-0e57-4d97-bc22-c8b4e2a81126 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.718346] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fa1511-90c7-4d93-813b-5fa05cfe25b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.724116] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Created directory with path [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 541.724364] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Fetch image to [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 541.724645] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloading image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk on the data store datastore1 {{(pid=61964) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 541.750977] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edff80a-02fd-48b7-ae1d-3c7cdba52cc6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.754746] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dabb128-53ce-45df-8fa0-7463fd8fe8f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.763605] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5f47f5-0321-4d6e-8f2e-8d4085561c9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.767527] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29dcd906-92ea-406e-beeb-e1b88513c90c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.789363] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b99ccf-68e3-4459-95df-20beb0197821 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.794101] env[61964]: DEBUG nova.compute.provider_tree [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.824592] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161ddef8-2671-4eba-924a-19243f6327b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.833149] env[61964]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-d2f6483e-accc-4c3c-9b8f-49dbcf2310b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.866405] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloading image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to the data store datastore1 {{(pid=61964) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 541.870995] env[61964]: INFO nova.scheduler.client.report [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Deleted allocations for instance e23a7d59-0516-46b9-ab6d-e4c313ad758a [ 541.877964] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.938915] env[61964]: DEBUG oslo_vmware.rw_handles [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 542.000522] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 542.130125] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.297312] env[61964]: DEBUG nova.scheduler.client.report [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 542.384732] env[61964]: DEBUG oslo_concurrency.lockutils [None req-74dfa858-cf0b-4425-a0e5-5db3f0cd9fea tempest-ServerActionsTestJSON-438603653 tempest-ServerActionsTestJSON-438603653-project-member] Lock "e23a7d59-0516-46b9-ab6d-e4c313ad758a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.983s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.633065] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Releasing lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.634015] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 542.634015] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.634175] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] Acquired lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.634329] env[61964]: DEBUG nova.network.neutron [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Refreshing network info cache for port f4628cb6-2328-403d-aa29-6dc2c4191ef1 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.635508] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17f7cbbf-8f36-432e-934b-3413b6562324 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.645898] env[61964]: DEBUG oslo_vmware.rw_handles [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 542.646104] env[61964]: DEBUG oslo_vmware.rw_handles [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 542.653782] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8d972f-a4f9-44c5-9ebe-a9d8e8a33f9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.666252] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Successfully created port: 5109dac7-31e2-4a20-be1a-aa626808b400 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.680189] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51b42c8a-f780-4725-85ab-3fad0d621dfd could not be found. [ 542.680442] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.680623] env[61964]: INFO nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 542.680873] env[61964]: DEBUG oslo.service.loopingcall [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.681111] env[61964]: DEBUG nova.compute.manager [-] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 542.681204] env[61964]: DEBUG nova.network.neutron [-] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.714579] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Downloaded image file data 5f28385d-6ea0-420d-8a26-4cb693714c14 to vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk on the data store datastore1 {{(pid=61964) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 542.719784] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 542.719784] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copying Virtual Disk [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk to [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 542.719784] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6e5ddfd-884c-4faa-a84d-15112bc416b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.731985] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 542.731985] env[61964]: value = "task-1040513" [ 542.731985] env[61964]: _type = "Task" [ 542.731985] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.742139] env[61964]: DEBUG nova.network.neutron [-] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.748753] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.805020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.868s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.805020] env[61964]: ERROR nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Traceback (most recent call last): [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.driver.spawn(context, instance, image_meta, [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.805020] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] vm_ref = self.build_virtual_machine(instance, [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] for vif in network_info: [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self._sync_wrapper(fn, *args, **kwargs) [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.wait() [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self[:] = self._gt.wait() [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self._exit_event.wait() [ 542.805541] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] result = hub.switch() [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return self.greenlet.switch() [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] result = function(*args, **kwargs) [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] return func(*args, **kwargs) [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise e [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] nwinfo = self.network_api.allocate_for_instance( [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.805903] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] created_port_ids = self._update_ports_for_instance( [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] with excutils.save_and_reraise_exception(): [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] self.force_reraise() [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise self.value [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] updated_port = self._update_port( [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] _ensure_no_port_binding_failure(port) [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.806240] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] raise exception.PortBindingFailed(port_id=port['id']) [ 542.806576] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] nova.exception.PortBindingFailed: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. [ 542.806576] env[61964]: ERROR nova.compute.manager [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] [ 542.806576] env[61964]: DEBUG nova.compute.utils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 542.807461] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.129s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.809853] env[61964]: INFO nova.compute.claims [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.812575] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Build of instance e5ee1938-27ad-4189-b9f0-9c9821e0ffc8 was re-scheduled: Binding failed for port 326413a6-397d-409b-881a-d0dffec7e92d, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 542.813044] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 542.813383] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquiring lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.813598] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Acquired lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.813985] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.886798] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 543.011042] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 543.046972] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 543.047213] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.047730] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 543.047944] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.048102] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 543.048250] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 543.048458] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 543.048611] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 543.048774] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 543.048937] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 543.049127] env[61964]: DEBUG nova.virt.hardware [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 543.050503] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691b771d-7fae-4cad-97ed-daa7a6ebea25 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.060055] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a635fe79-493f-47ac-844c-161871b5e783 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.195022] env[61964]: DEBUG nova.network.neutron [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.249018] env[61964]: DEBUG nova.network.neutron [-] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.251902] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040513, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.357180] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.420843] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.533123] env[61964]: DEBUG nova.network.neutron [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.624619] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.748779] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652382} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.749056] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copied Virtual Disk [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk to [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 543.749605] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14/tmp-sparse.vmdk {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 543.749605] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1891558e-21f3-4bc2-878e-efef6ce6c0c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.754362] env[61964]: INFO nova.compute.manager [-] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Took 1.07 seconds to deallocate network for instance. [ 543.758899] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 543.758899] env[61964]: value = "task-1040514" [ 543.758899] env[61964]: _type = "Task" [ 543.758899] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.759774] env[61964]: DEBUG nova.compute.claims [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.759993] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.772841] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.037477] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] Releasing lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.037619] env[61964]: DEBUG nova.compute.manager [req-f1f2ceb3-9463-4dba-9ce1-5b378cff4413 req-ecee0a19-9e0a-4940-a128-d6b1aa1290eb service nova] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Received event network-vif-deleted-f4628cb6-2328-403d-aa29-6dc2c4191ef1 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 544.060523] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02da831-5e58-4767-bd8d-3514969a21ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.069523] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c8562d-85fd-406c-9f26-e10d80cf5d1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.104760] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb146e3-a21f-4fcc-b608-6bb56dc95755 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.114419] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20722bc-7404-4df3-91c4-d052d0d911a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.129819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Releasing lock "refresh_cache-e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.130085] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 544.130249] env[61964]: DEBUG nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 544.130467] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 544.132392] env[61964]: DEBUG nova.compute.provider_tree [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.166976] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.273162] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025439} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.273502] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 544.274566] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Moving file from [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068/5f28385d-6ea0-420d-8a26-4cb693714c14 to [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14. {{(pid=61964) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 544.274566] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e0ac84d7-437c-4392-9d10-b9a24a604ec5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.281765] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 544.281765] env[61964]: value = "task-1040515" [ 544.281765] env[61964]: _type = "Task" [ 544.281765] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.291240] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040515, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.638039] env[61964]: DEBUG nova.scheduler.client.report [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 544.670908] env[61964]: DEBUG nova.network.neutron [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.793867] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040515, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.041248} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.795215] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] File moved {{(pid=61964) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 544.795561] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Cleaning up location [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 544.795987] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore1] vmware_temp/2b382a30-fb1d-4600-b26a-4959a650c068 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 544.798019] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-093c42e9-5950-4a25-8cc2-74b0ebf50baa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.805114] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 544.805114] env[61964]: value = "task-1040516" [ 544.805114] env[61964]: _type = "Task" [ 544.805114] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.815777] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.096992] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.097275] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.142956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.143766] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 545.147412] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.446s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.172945] env[61964]: INFO nova.compute.manager [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] [instance: e5ee1938-27ad-4189-b9f0-9c9821e0ffc8] Took 1.04 seconds to deallocate network for instance. [ 545.323990] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026618} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.326692] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 545.326692] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef7215b2-3770-404d-bf0c-2484d0d97a03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.336766] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 545.336766] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52873656-0b8c-945d-a3b3-b161666dea45" [ 545.336766] env[61964]: _type = "Task" [ 545.336766] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.353689] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52873656-0b8c-945d-a3b3-b161666dea45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.654695] env[61964]: DEBUG nova.compute.utils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 545.658780] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 545.658956] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 545.756899] env[61964]: DEBUG nova.policy [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9e9a88a96aa47468a951dfaaa607fba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c2af79e2f474958b091710801a95749', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 545.851395] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52873656-0b8c-945d-a3b3-b161666dea45, 'name': SearchDatastore_Task, 'duration_secs': 0.015897} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.854411] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.854534] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 545.854975] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c643f65-7722-4062-b5d8-08cbc80c0984 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.862962] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 545.862962] env[61964]: value = "task-1040517" [ 545.862962] env[61964]: _type = "Task" [ 545.862962] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.876929] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.926438] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8631a6f8-58ec-4107-93ab-75b6d4269fbf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.930350] env[61964]: ERROR nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 545.930350] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.930350] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.930350] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.930350] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.930350] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.930350] env[61964]: ERROR nova.compute.manager raise self.value [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.930350] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.930350] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.930350] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.930865] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.930865] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.930865] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 545.930865] env[61964]: ERROR nova.compute.manager [ 545.930865] env[61964]: Traceback (most recent call last): [ 545.930865] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.930865] env[61964]: listener.cb(fileno) [ 545.930865] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.930865] env[61964]: result = function(*args, **kwargs) [ 545.930865] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.930865] env[61964]: return func(*args, **kwargs) [ 545.930865] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.930865] env[61964]: raise e [ 545.930865] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.930865] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 545.930865] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.930865] env[61964]: created_port_ids = self._update_ports_for_instance( [ 545.930865] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.930865] env[61964]: with excutils.save_and_reraise_exception(): [ 545.930865] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.930865] env[61964]: self.force_reraise() [ 545.930865] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.930865] env[61964]: raise self.value [ 545.930865] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.930865] env[61964]: updated_port = self._update_port( [ 545.930865] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.930865] env[61964]: _ensure_no_port_binding_failure(port) [ 545.930865] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.930865] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.931738] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 545.931738] env[61964]: Removing descriptor: 17 [ 545.931738] env[61964]: ERROR nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Traceback (most recent call last): [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] yield resources [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.driver.spawn(context, instance, image_meta, [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.931738] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] vm_ref = self.build_virtual_machine(instance, [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] for vif in network_info: [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self._sync_wrapper(fn, *args, **kwargs) [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.wait() [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self[:] = self._gt.wait() [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self._exit_event.wait() [ 545.932118] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] result = hub.switch() [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self.greenlet.switch() [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] result = function(*args, **kwargs) [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return func(*args, **kwargs) [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise e [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] nwinfo = self.network_api.allocate_for_instance( [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.932538] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] created_port_ids = self._update_ports_for_instance( [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] with excutils.save_and_reraise_exception(): [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.force_reraise() [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise self.value [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] updated_port = self._update_port( [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] _ensure_no_port_binding_failure(port) [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.933012] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise exception.PortBindingFailed(port_id=port['id']) [ 545.933452] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 545.933452] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] [ 545.933452] env[61964]: INFO nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Terminating instance [ 545.940475] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610ac772-b543-4bf7-86ee-733fe6ad40f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.977239] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff11f154-072b-44de-a5d8-caa2176c6685 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.985322] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5876657a-9f27-4468-9778-858dece91b91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.001523] env[61964]: DEBUG nova.compute.provider_tree [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.162363] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 546.223734] env[61964]: INFO nova.scheduler.client.report [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Deleted allocations for instance e5ee1938-27ad-4189-b9f0-9c9821e0ffc8 [ 546.383626] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040517, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.442281] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.442456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.442650] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.504046] env[61964]: DEBUG nova.scheduler.client.report [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 546.531494] env[61964]: DEBUG nova.compute.manager [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Received event network-changed-5109dac7-31e2-4a20-be1a-aa626808b400 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 546.531688] env[61964]: DEBUG nova.compute.manager [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Refreshing instance network info cache due to event network-changed-5109dac7-31e2-4a20-be1a-aa626808b400. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 546.531878] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] Acquiring lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.545160] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Successfully created port: 5a6b83b1-d922-4373-8923-2f297a695e94 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.734654] env[61964]: DEBUG oslo_concurrency.lockutils [None req-31a7cd81-5f8f-4ef3-a271-c414779075d7 tempest-ServerDiagnosticsTest-1683969254 tempest-ServerDiagnosticsTest-1683969254-project-member] Lock "e5ee1938-27ad-4189-b9f0-9c9821e0ffc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.458s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.880542] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522326} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.880904] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 546.882099] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 546.882543] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82ef1e13-e721-4cbc-add4-5b6a9cf2eacc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.893733] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 546.893733] env[61964]: value = "task-1040518" [ 546.893733] env[61964]: _type = "Task" [ 546.893733] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.904588] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.985580] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.018135] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.018135] env[61964]: ERROR nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Traceback (most recent call last): [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.driver.spawn(context, instance, image_meta, [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.018135] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] vm_ref = self.build_virtual_machine(instance, [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] for vif in network_info: [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self._sync_wrapper(fn, *args, **kwargs) [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.wait() [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self[:] = self._gt.wait() [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self._exit_event.wait() [ 547.018622] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] result = hub.switch() [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return self.greenlet.switch() [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] result = function(*args, **kwargs) [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] return func(*args, **kwargs) [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise e [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] nwinfo = self.network_api.allocate_for_instance( [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.018942] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] created_port_ids = self._update_ports_for_instance( [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] with excutils.save_and_reraise_exception(): [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] self.force_reraise() [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise self.value [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] updated_port = self._update_port( [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] _ensure_no_port_binding_failure(port) [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.019282] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] raise exception.PortBindingFailed(port_id=port['id']) [ 547.019559] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] nova.exception.PortBindingFailed: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. [ 547.019559] env[61964]: ERROR nova.compute.manager [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] [ 547.019559] env[61964]: DEBUG nova.compute.utils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.019559] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Build of instance 7ec4c677-9166-4b3c-ade7-548bee8e245b was re-scheduled: Binding failed for port e7818c75-73c2-4809-9d4b-6c43293258c4, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 547.019692] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 547.020733] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.020733] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquired lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.020733] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.021848] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.167s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.074418] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquiring lock "08bd2fb8-fc8d-440c-9369-196605a894e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.074679] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "08bd2fb8-fc8d-440c-9369-196605a894e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.179103] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 547.207909] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 547.208251] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 547.208566] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 547.208662] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 547.208810] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 547.209880] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 547.209880] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 547.209880] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 547.209880] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 547.214152] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 547.214152] env[61964]: DEBUG nova.virt.hardware [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 547.214152] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c80770-1604-4f31-a5b3-c39c5a5fbce5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.225973] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f398cd56-b16a-4c88-a18a-791e0ed10f98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.252726] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 547.352050] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.405686] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066741} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.409037] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 547.409037] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7aac36-ce38-4427-9d66-1a47bc01a6a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.440079] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 547.441506] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaa98149-ac5b-41c5-939a-7363ae5dc8f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.462341] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 547.462341] env[61964]: value = "task-1040519" [ 547.462341] env[61964]: _type = "Task" [ 547.462341] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.470741] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040519, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.575791] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.773394] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.780894] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.812511] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a07deb8-6dd8-423a-be1d-f224bfb9868b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.821689] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de3438e-5d47-4ece-920b-77a56f08faba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.873493] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.873493] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.874843] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 547.874843] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] Acquired lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.875457] env[61964]: DEBUG nova.network.neutron [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Refreshing network info cache for port 5109dac7-31e2-4a20-be1a-aa626808b400 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 547.876537] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-351b1914-ff1c-4f21-b00c-2198d787cb38 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.880528] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1da3a2-6ce8-4b38-974d-190718a18ec3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.890709] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquiring lock "e95ab551-63c1-4fa2-b615-5c6a543c39c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.891048] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "e95ab551-63c1-4fa2-b615-5c6a543c39c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.896438] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde37faa-e658-4649-aba0-bf5b3fb41461 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.906146] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c51a2e-3afc-430d-82e7-43d1d354fadd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.928616] env[61964]: DEBUG nova.compute.provider_tree [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.934873] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee058557-cace-4e32-bdf9-bd4af7fe29ff could not be found. [ 547.935099] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.935284] env[61964]: INFO nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Took 0.06 seconds to destroy the instance on the hypervisor. [ 547.935524] env[61964]: DEBUG oslo.service.loopingcall [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.936071] env[61964]: DEBUG nova.compute.manager [-] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.936216] env[61964]: DEBUG nova.network.neutron [-] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.971353] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040519, 'name': ReconfigVM_Task, 'duration_secs': 0.28491} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.971658] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 547.972298] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8fdd25b-24b0-4b36-b3f6-38f540f7b0d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.978683] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 547.978683] env[61964]: value = "task-1040520" [ 547.978683] env[61964]: _type = "Task" [ 547.978683] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.982281] env[61964]: DEBUG nova.network.neutron [-] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.989155] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040520, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.277560] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Releasing lock "refresh_cache-7ec4c677-9166-4b3c-ade7-548bee8e245b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.278629] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 548.278729] env[61964]: DEBUG nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 548.280974] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.316838] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.423555] env[61964]: DEBUG nova.network.neutron [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.439653] env[61964]: DEBUG nova.scheduler.client.report [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 548.489949] env[61964]: DEBUG nova.network.neutron [-] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.491159] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040520, 'name': Rename_Task, 'duration_secs': 0.140767} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.491423] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 548.491873] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f5f8037-1f97-4ddb-9703-1a45deab5b67 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.497876] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 548.497876] env[61964]: value = "task-1040521" [ 548.497876] env[61964]: _type = "Task" [ 548.497876] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.507122] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.572152] env[61964]: DEBUG nova.network.neutron [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.819056] env[61964]: DEBUG nova.network.neutron [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.949326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.949895] env[61964]: ERROR nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Traceback (most recent call last): [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.driver.spawn(context, instance, image_meta, [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] vm_ref = self.build_virtual_machine(instance, [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.949895] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] for vif in network_info: [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self._sync_wrapper(fn, *args, **kwargs) [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.wait() [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self[:] = self._gt.wait() [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self._exit_event.wait() [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] result = hub.switch() [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.950292] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return self.greenlet.switch() [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] result = function(*args, **kwargs) [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] return func(*args, **kwargs) [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise e [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] nwinfo = self.network_api.allocate_for_instance( [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] created_port_ids = self._update_ports_for_instance( [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] with excutils.save_and_reraise_exception(): [ 548.950654] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] self.force_reraise() [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise self.value [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] updated_port = self._update_port( [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] _ensure_no_port_binding_failure(port) [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] raise exception.PortBindingFailed(port_id=port['id']) [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] nova.exception.PortBindingFailed: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. [ 548.950994] env[61964]: ERROR nova.compute.manager [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] [ 548.951308] env[61964]: DEBUG nova.compute.utils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.951711] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.921s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.954687] env[61964]: INFO nova.compute.claims [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.962327] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Build of instance 526c58ea-0813-4d49-99cf-11d32bfcca30 was re-scheduled: Binding failed for port ab3a7988-20c1-486c-8b69-9c47f0780662, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 548.962327] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 548.962327] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquiring lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.962327] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Acquired lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.962595] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.993071] env[61964]: INFO nova.compute.manager [-] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Took 1.06 seconds to deallocate network for instance. [ 548.997910] env[61964]: DEBUG nova.compute.claims [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.997910] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.012741] env[61964]: DEBUG oslo_vmware.api [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040521, 'name': PowerOnVM_Task, 'duration_secs': 0.414972} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.013604] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 549.013788] env[61964]: DEBUG nova.compute.manager [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 549.014599] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ea79b1-039d-46de-9179-fa46c49ece0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.078518] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c095e58-4ceb-4411-b8d7-94b3865d9cae req-7de680df-dc4c-44ac-bb08-cd59e86cb8fa service nova] Releasing lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.100526] env[61964]: DEBUG nova.compute.manager [req-94ac539c-f113-4b1c-a766-58a0da53a67e req-520f910a-6631-4f0e-9dd4-bacf0694e618 service nova] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Received event network-vif-deleted-5109dac7-31e2-4a20-be1a-aa626808b400 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 549.327206] env[61964]: INFO nova.compute.manager [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 7ec4c677-9166-4b3c-ade7-548bee8e245b] Took 1.05 seconds to deallocate network for instance. [ 549.499365] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.540508] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.783233] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.019393] env[61964]: ERROR nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 550.019393] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.019393] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.019393] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.019393] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.019393] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.019393] env[61964]: ERROR nova.compute.manager raise self.value [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.019393] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.019393] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.019393] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.020572] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.020572] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.020572] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 550.020572] env[61964]: ERROR nova.compute.manager [ 550.020572] env[61964]: Traceback (most recent call last): [ 550.020572] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.020572] env[61964]: listener.cb(fileno) [ 550.020572] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.020572] env[61964]: result = function(*args, **kwargs) [ 550.020572] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.020572] env[61964]: return func(*args, **kwargs) [ 550.020572] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.020572] env[61964]: raise e [ 550.020572] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.020572] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 550.020572] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.020572] env[61964]: created_port_ids = self._update_ports_for_instance( [ 550.020572] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.020572] env[61964]: with excutils.save_and_reraise_exception(): [ 550.020572] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.020572] env[61964]: self.force_reraise() [ 550.020572] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.020572] env[61964]: raise self.value [ 550.020572] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.020572] env[61964]: updated_port = self._update_port( [ 550.020572] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.020572] env[61964]: _ensure_no_port_binding_failure(port) [ 550.020572] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.020572] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.021400] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 550.021400] env[61964]: Removing descriptor: 15 [ 550.021400] env[61964]: ERROR nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Traceback (most recent call last): [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] yield resources [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.driver.spawn(context, instance, image_meta, [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.021400] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] vm_ref = self.build_virtual_machine(instance, [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] for vif in network_info: [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self._sync_wrapper(fn, *args, **kwargs) [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.wait() [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self[:] = self._gt.wait() [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self._exit_event.wait() [ 550.021972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] result = hub.switch() [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self.greenlet.switch() [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] result = function(*args, **kwargs) [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return func(*args, **kwargs) [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise e [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] nwinfo = self.network_api.allocate_for_instance( [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.022327] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] created_port_ids = self._update_ports_for_instance( [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] with excutils.save_and_reraise_exception(): [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.force_reraise() [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise self.value [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] updated_port = self._update_port( [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] _ensure_no_port_binding_failure(port) [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.022667] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise exception.PortBindingFailed(port_id=port['id']) [ 550.022972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 550.022972] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] [ 550.022972] env[61964]: INFO nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Terminating instance [ 550.231437] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b996b301-4640-4bb0-89e2-b4186a28a4ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.240943] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b847fe44-ad04-46f1-8625-198cb492b931 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.278248] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b612811-ea33-4ed0-89c4-62347971a990 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.286231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adba6bf-0d6e-4c05-ba11-7f5e025c8f93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.292517] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Releasing lock "refresh_cache-526c58ea-0813-4d49-99cf-11d32bfcca30" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.293243] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 550.293243] env[61964]: DEBUG nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 550.293415] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.305401] env[61964]: DEBUG nova.compute.provider_tree [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.330859] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.361807] env[61964]: INFO nova.scheduler.client.report [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Deleted allocations for instance 7ec4c677-9166-4b3c-ade7-548bee8e245b [ 550.525830] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquiring lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.526068] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquired lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.526221] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 550.808235] env[61964]: DEBUG nova.scheduler.client.report [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 550.843366] env[61964]: DEBUG nova.network.neutron [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.871570] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f3861472-cb74-4706-981e-3a74ff2b8c0c tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "7ec4c677-9166-4b3c-ade7-548bee8e245b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.493s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.066373] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.318355] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.318355] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 551.328020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.506s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.349310] env[61964]: INFO nova.compute.manager [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] [instance: 526c58ea-0813-4d49-99cf-11d32bfcca30] Took 1.05 seconds to deallocate network for instance. [ 551.356227] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.385870] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 551.502353] env[61964]: DEBUG nova.compute.manager [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Received event network-changed-5a6b83b1-d922-4373-8923-2f297a695e94 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 551.502559] env[61964]: DEBUG nova.compute.manager [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Refreshing instance network info cache due to event network-changed-5a6b83b1-d922-4373-8923-2f297a695e94. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 551.505767] env[61964]: DEBUG oslo_concurrency.lockutils [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] Acquiring lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.505767] env[61964]: INFO nova.compute.manager [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Rebuilding instance [ 551.597032] env[61964]: DEBUG nova.compute.manager [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 551.597032] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296330e3-cf2b-428e-8434-e26a6bd2ccdb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.831070] env[61964]: DEBUG nova.compute.utils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.839837] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 551.839837] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 551.860563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Releasing lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.862700] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 551.862903] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 551.863593] env[61964]: DEBUG oslo_concurrency.lockutils [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] Acquired lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.863593] env[61964]: DEBUG nova.network.neutron [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Refreshing network info cache for port 5a6b83b1-d922-4373-8923-2f297a695e94 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 551.864487] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7dc8df13-b252-470e-83ba-1d005abf72ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.887768] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45bc1e7-30c5-40c2-ac24-22ea24ef5742 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.921417] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9e79aac-19af-4339-9b06-5907d2ac716c could not be found. [ 551.921663] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 551.921854] env[61964]: INFO nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Took 0.06 seconds to destroy the instance on the hypervisor. [ 551.922126] env[61964]: DEBUG oslo.service.loopingcall [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.922353] env[61964]: DEBUG nova.compute.manager [-] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.922451] env[61964]: DEBUG nova.network.neutron [-] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.926730] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.981034] env[61964]: DEBUG nova.policy [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbcb6948440e47c1b88eb0002d6931ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '486fd4a100d14be6839a137f480069e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.983730] env[61964]: DEBUG nova.network.neutron [-] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.339714] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 552.392021] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 4e14a519-b543-4a18-9e1d-7df16abf5345 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.413196] env[61964]: INFO nova.scheduler.client.report [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Deleted allocations for instance 526c58ea-0813-4d49-99cf-11d32bfcca30 [ 552.449535] env[61964]: DEBUG nova.network.neutron [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.487762] env[61964]: DEBUG nova.network.neutron [-] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.614404] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 552.614722] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf103d72-d5ed-4063-b171-11f7c2c1409a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.625480] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 552.625480] env[61964]: value = "task-1040522" [ 552.625480] env[61964]: _type = "Task" [ 552.625480] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.637983] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.813406] env[61964]: DEBUG nova.network.neutron [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.898350] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 526c58ea-0813-4d49-99cf-11d32bfcca30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.898478] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6317e489-9347-46b7-989a-f543af500b39 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.898656] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 894ad0a3-851a-4b5e-82b0-050cbbf264f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.898747] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 51b42c8a-f780-4725-85ab-3fad0d621dfd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.898829] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance ee058557-cace-4e32-bdf9-bd4af7fe29ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.898994] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance c9e79aac-19af-4339-9b06-5907d2ac716c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.899054] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 07771b38-3dc9-4608-8bea-48971bbc9651 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 552.927278] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d738b72-5bbf-465b-a758-34f12263c7ce tempest-ImagesOneServerNegativeTestJSON-1439203133 tempest-ImagesOneServerNegativeTestJSON-1439203133-project-member] Lock "526c58ea-0813-4d49-99cf-11d32bfcca30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.138s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.944805] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Successfully created port: 24baf232-a5bb-4017-a72b-2545bf292cce {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.994040] env[61964]: INFO nova.compute.manager [-] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Took 1.07 seconds to deallocate network for instance. [ 552.997689] env[61964]: DEBUG nova.compute.claims [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.997960] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.140277] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040522, 'name': PowerOffVM_Task, 'duration_secs': 0.147591} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.140277] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 553.140502] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 553.141294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068ee183-cef2-4e0f-af4d-0bac8357e036 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.149826] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 553.150128] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02b648a2-d7cb-4dad-b68b-2b1f231e874a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.172896] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 553.173191] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 553.173411] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Deleting the datastore file [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.173670] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4600c28-85dd-48ab-9260-0fe870ebca61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.180262] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 553.180262] env[61964]: value = "task-1040524" [ 553.180262] env[61964]: _type = "Task" [ 553.180262] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.192012] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.317743] env[61964]: DEBUG oslo_concurrency.lockutils [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] Releasing lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.317743] env[61964]: DEBUG nova.compute.manager [req-241b6848-ba0e-4baa-8f11-c90461f24761 req-47c7619c-1e1c-4d10-940d-00548e0032bd service nova] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Received event network-vif-deleted-5a6b83b1-d922-4373-8923-2f297a695e94 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 553.352151] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 553.380176] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 553.381420] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.381593] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 553.381771] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.381914] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 553.382069] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 553.382339] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 553.382506] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 553.382682] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 553.382834] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 553.383989] env[61964]: DEBUG nova.virt.hardware [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 553.383989] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40aaa6b3-9099-459a-be8b-5061005f01f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.392770] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b91da7-97df-4399-a12a-c7f1ee981483 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.407674] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.438955] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 553.692735] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113409} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.692997] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 553.693466] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 553.693466] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 553.910245] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance c93818ca-3f7c-4cd2-b8c4-d7714c4404a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.977677] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.373509] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquiring lock "19a6958e-4887-459b-8bf7-dd4cbbedb700" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.373846] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "19a6958e-4887-459b-8bf7-dd4cbbedb700" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.416788] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 84cda217-8acd-4fd9-80ac-cb83edcc1347 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.635654] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "6c45bf0b-17c2-4065-94d9-52ee7b9af151" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.636095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "6c45bf0b-17c2-4065-94d9-52ee7b9af151" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.736795] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 554.737052] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 554.737215] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 554.737390] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 554.737534] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 554.737677] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 554.737934] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 554.741317] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 554.741649] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 554.741760] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 554.741941] env[61964]: DEBUG nova.virt.hardware [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 554.742821] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894f3c07-4920-48b6-a650-145e484dbfa4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.753443] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1374ed2-9054-423e-b13e-f519a16fd9d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.772765] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 554.779817] env[61964]: DEBUG oslo.service.loopingcall [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.779817] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 554.780015] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c78d268c-cdba-43f2-9754-ec3f4c82f9f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.800061] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 554.800061] env[61964]: value = "task-1040525" [ 554.800061] env[61964]: _type = "Task" [ 554.800061] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.811840] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040525, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.920593] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.314208] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040525, 'name': CreateVM_Task, 'duration_secs': 0.284408} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.314208] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 555.314208] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.314208] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.317586] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 555.317586] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-feb53c1f-f090-4df4-8460-ea5e9584c02e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.320344] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 555.320344] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4c9cc-8023-6cb9-579e-a4a3aa4d2df4" [ 555.320344] env[61964]: _type = "Task" [ 555.320344] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.332520] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4c9cc-8023-6cb9-579e-a4a3aa4d2df4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.427382] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 08bd2fb8-fc8d-440c-9369-196605a894e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.835328] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4c9cc-8023-6cb9-579e-a4a3aa4d2df4, 'name': SearchDatastore_Task, 'duration_secs': 0.008151} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.835649] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.835884] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 555.836170] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.836320] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.840248] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 555.840248] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-846ae73e-7c54-42a4-87f1-1948d0095a7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.847217] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 555.847417] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 555.848221] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e6f0ce7-e21e-48e2-b232-299e93ca3087 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.854693] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 555.854693] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52383f9e-e05d-7328-5b07-8bb32b34ea87" [ 555.854693] env[61964]: _type = "Task" [ 555.854693] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.866777] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52383f9e-e05d-7328-5b07-8bb32b34ea87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.932430] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance e95ab551-63c1-4fa2-b615-5c6a543c39c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.933444] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 555.934643] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 556.186125] env[61964]: ERROR nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 556.186125] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.186125] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.186125] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.186125] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.186125] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.186125] env[61964]: ERROR nova.compute.manager raise self.value [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.186125] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.186125] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.186125] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.188022] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.188022] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.188022] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 556.188022] env[61964]: ERROR nova.compute.manager [ 556.188022] env[61964]: Traceback (most recent call last): [ 556.188022] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.188022] env[61964]: listener.cb(fileno) [ 556.188022] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.188022] env[61964]: result = function(*args, **kwargs) [ 556.188022] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.188022] env[61964]: return func(*args, **kwargs) [ 556.188022] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.188022] env[61964]: raise e [ 556.188022] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.188022] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 556.188022] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.188022] env[61964]: created_port_ids = self._update_ports_for_instance( [ 556.188022] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.188022] env[61964]: with excutils.save_and_reraise_exception(): [ 556.188022] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.188022] env[61964]: self.force_reraise() [ 556.188022] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.188022] env[61964]: raise self.value [ 556.188022] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.188022] env[61964]: updated_port = self._update_port( [ 556.188022] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.188022] env[61964]: _ensure_no_port_binding_failure(port) [ 556.188022] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.188022] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.188912] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 556.188912] env[61964]: Removing descriptor: 15 [ 556.188912] env[61964]: ERROR nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Traceback (most recent call last): [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] yield resources [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.driver.spawn(context, instance, image_meta, [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.188912] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] vm_ref = self.build_virtual_machine(instance, [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] for vif in network_info: [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self._sync_wrapper(fn, *args, **kwargs) [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.wait() [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self[:] = self._gt.wait() [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self._exit_event.wait() [ 556.189258] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] result = hub.switch() [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self.greenlet.switch() [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] result = function(*args, **kwargs) [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return func(*args, **kwargs) [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise e [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] nwinfo = self.network_api.allocate_for_instance( [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.189615] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] created_port_ids = self._update_ports_for_instance( [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] with excutils.save_and_reraise_exception(): [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.force_reraise() [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise self.value [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] updated_port = self._update_port( [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] _ensure_no_port_binding_failure(port) [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.189976] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise exception.PortBindingFailed(port_id=port['id']) [ 556.190316] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 556.190316] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] [ 556.190316] env[61964]: INFO nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Terminating instance [ 556.199044] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquiring lock "4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.199290] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.327124] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065868e7-fad0-47ab-9162-481e212b4ab9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.336623] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d8ad27-d64e-43b8-8fcc-ecb7be8f80d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.374388] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63beec7e-2f87-4926-acb0-3f9d4aae86b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.383689] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52383f9e-e05d-7328-5b07-8bb32b34ea87, 'name': SearchDatastore_Task, 'duration_secs': 0.008235} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.385424] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72fe07d9-46a5-48c5-b9b7-b49a4a47bafd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.388665] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4797536-52e4-4c6c-af3d-974138116196 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.402054] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.405782] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 556.405782] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52abf7d2-9956-cd0e-6817-778cb51ed9c4" [ 556.405782] env[61964]: _type = "Task" [ 556.405782] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.412789] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52abf7d2-9956-cd0e-6817-778cb51ed9c4, 'name': SearchDatastore_Task, 'duration_secs': 0.008853} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.413069] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.413565] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 556.413684] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6f176a0-eaa3-401f-82c3-3f242f519e11 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.420037] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 556.420037] env[61964]: value = "task-1040526" [ 556.420037] env[61964]: _type = "Task" [ 556.420037] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.430352] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.466774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "669fe2d9-cb37-4757-a199-18877cc6149e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.466774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "669fe2d9-cb37-4757-a199-18877cc6149e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.626041] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "1167bfd6-ba4f-4b19-bea6-e934bc61829d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.626412] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "1167bfd6-ba4f-4b19-bea6-e934bc61829d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.697527] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.697745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquired lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.697953] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.724535] env[61964]: DEBUG nova.compute.manager [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Received event network-changed-24baf232-a5bb-4017-a72b-2545bf292cce {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 556.724535] env[61964]: DEBUG nova.compute.manager [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Refreshing instance network info cache due to event network-changed-24baf232-a5bb-4017-a72b-2545bf292cce. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 556.724535] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] Acquiring lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.906334] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 556.931512] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489975} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.931781] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 556.932024] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 556.932308] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7732cb6-b112-4f01-be0d-0f535a2a63d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.941411] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 556.941411] env[61964]: value = "task-1040527" [ 556.941411] env[61964]: _type = "Task" [ 556.941411] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.953126] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.227844] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.401614] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.411101] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 557.411564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.089s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.411824] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.049s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.454243] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065469} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.454811] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 557.457173] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e943c467-f7cd-4813-bc86-deba687a80c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.481453] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 557.481824] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c69a42dc-789e-4508-9a16-0f6839b7d7d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.502912] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 557.502912] env[61964]: value = "task-1040528" [ 557.502912] env[61964]: _type = "Task" [ 557.502912] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.511911] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.904378] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Releasing lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.904902] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 557.905088] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 557.905464] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] Acquired lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.905654] env[61964]: DEBUG nova.network.neutron [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Refreshing network info cache for port 24baf232-a5bb-4017-a72b-2545bf292cce {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 557.906811] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05783ef3-fc11-473f-b977-627ea92bdf15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.927260] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69404e6c-c238-4466-925e-c15335bb7825 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.958819] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 07771b38-3dc9-4608-8bea-48971bbc9651 could not be found. [ 557.959055] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 557.959241] env[61964]: INFO nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Took 0.05 seconds to destroy the instance on the hypervisor. [ 557.959493] env[61964]: DEBUG oslo.service.loopingcall [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.962115] env[61964]: DEBUG nova.compute.manager [-] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 557.962453] env[61964]: DEBUG nova.network.neutron [-] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.002438] env[61964]: DEBUG nova.network.neutron [-] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.013358] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.262690] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee19cd2-cbd8-44b1-b4a6-18f27cd342f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.271201] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554158f6-8a94-4429-a979-0e57b6b9c2d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.305485] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ed09d5-8300-45bb-b0b8-736e27ab6c25 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.313270] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59e642a-e7eb-4dc7-9d7c-de7b5037d6a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.330778] env[61964]: DEBUG nova.compute.provider_tree [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.455710] env[61964]: DEBUG nova.network.neutron [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.511237] env[61964]: DEBUG nova.network.neutron [-] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.520025] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.665483] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "64818b22-3bd5-473e-8474-40bbb40ab491" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.665722] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "64818b22-3bd5-473e-8474-40bbb40ab491" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.810398] env[61964]: DEBUG nova.network.neutron [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.832099] env[61964]: DEBUG nova.scheduler.client.report [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 559.016254] env[61964]: INFO nova.compute.manager [-] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Took 1.05 seconds to deallocate network for instance. [ 559.016715] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040528, 'name': ReconfigVM_Task, 'duration_secs': 1.264366} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.018649] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Reconfigured VM instance instance-00000002 to attach disk [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345/4e14a519-b543-4a18-9e1d-7df16abf5345.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 559.019608] env[61964]: DEBUG nova.compute.claims [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.019884] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.020332] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ebb6e43c-6ce3-46cc-906c-ae33fa86674a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.029609] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 559.029609] env[61964]: value = "task-1040529" [ 559.029609] env[61964]: _type = "Task" [ 559.029609] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.041788] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040529, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.314245] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ed29754-7695-49cd-9965-8a13c919f183 req-1e9b4e4c-35e4-4045-9379-c5a8e40d5613 service nova] Releasing lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.338377] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.340827] env[61964]: ERROR nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] Traceback (most recent call last): [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.driver.spawn(context, instance, image_meta, [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] vm_ref = self.build_virtual_machine(instance, [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.340827] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] for vif in network_info: [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self._sync_wrapper(fn, *args, **kwargs) [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.wait() [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self[:] = self._gt.wait() [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self._exit_event.wait() [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] result = hub.switch() [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.341376] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return self.greenlet.switch() [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] result = function(*args, **kwargs) [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] return func(*args, **kwargs) [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise e [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] nwinfo = self.network_api.allocate_for_instance( [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] created_port_ids = self._update_ports_for_instance( [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] with excutils.save_and_reraise_exception(): [ 559.341831] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] self.force_reraise() [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise self.value [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] updated_port = self._update_port( [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] _ensure_no_port_binding_failure(port) [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] raise exception.PortBindingFailed(port_id=port['id']) [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] nova.exception.PortBindingFailed: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. [ 559.342231] env[61964]: ERROR nova.compute.manager [instance: 6317e489-9347-46b7-989a-f543af500b39] [ 559.342560] env[61964]: DEBUG nova.compute.utils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.348315] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.871s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.350485] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Build of instance 6317e489-9347-46b7-989a-f543af500b39 was re-scheduled: Binding failed for port c42144e7-4206-4859-8d44-6f52bb282ffe, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 559.351259] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 559.352710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquiring lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.352710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Acquired lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.352710] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.467824] env[61964]: DEBUG nova.compute.manager [req-10a96310-3307-476d-8d1a-bc585f18c89f req-a82ef0a2-630a-44db-b65a-32bddf8cb53b service nova] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Received event network-vif-deleted-24baf232-a5bb-4017-a72b-2545bf292cce {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 559.546264] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040529, 'name': Rename_Task, 'duration_secs': 0.141052} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.546264] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 559.546264] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ea9e17b-9584-4373-ae77-01d1e24c3ee6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.551440] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Waiting for the task: (returnval){ [ 559.551440] env[61964]: value = "task-1040530" [ 559.551440] env[61964]: _type = "Task" [ 559.551440] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.560802] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.903404] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.085566] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040530, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.137582] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.277213] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bb81d6-604a-4720-a34e-9b1c3ac740d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.285509] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8746453-89e5-4556-87ea-9ea96b273df0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.331431] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635a1d9d-8247-4785-84d9-154d29afbd48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.340973] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab41d09b-6f40-4896-8857-b8026de88779 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.361432] env[61964]: DEBUG nova.compute.provider_tree [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.572425] env[61964]: DEBUG oslo_vmware.api [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Task: {'id': task-1040530, 'name': PowerOnVM_Task, 'duration_secs': 0.861392} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.572425] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 560.572425] env[61964]: DEBUG nova.compute.manager [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 560.572425] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a9d15d-9850-45b8-a56e-a37b6ba4065b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.641609] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Releasing lock "refresh_cache-6317e489-9347-46b7-989a-f543af500b39" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.641909] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 560.644283] env[61964]: DEBUG nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 560.644283] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.658767] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.868018] env[61964]: DEBUG nova.scheduler.client.report [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 561.096310] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.161928] env[61964]: DEBUG nova.network.neutron [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.373560] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.027s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.373752] env[61964]: ERROR nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Traceback (most recent call last): [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.driver.spawn(context, instance, image_meta, [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] vm_ref = self.build_virtual_machine(instance, [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.373752] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] for vif in network_info: [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self._sync_wrapper(fn, *args, **kwargs) [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.wait() [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self[:] = self._gt.wait() [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self._exit_event.wait() [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] result = hub.switch() [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.374157] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return self.greenlet.switch() [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] result = function(*args, **kwargs) [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] return func(*args, **kwargs) [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise e [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] nwinfo = self.network_api.allocate_for_instance( [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] created_port_ids = self._update_ports_for_instance( [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] with excutils.save_and_reraise_exception(): [ 561.374576] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] self.force_reraise() [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise self.value [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] updated_port = self._update_port( [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] _ensure_no_port_binding_failure(port) [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] raise exception.PortBindingFailed(port_id=port['id']) [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] nova.exception.PortBindingFailed: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. [ 561.374956] env[61964]: ERROR nova.compute.manager [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] [ 561.376119] env[61964]: DEBUG nova.compute.utils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.378589] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.957s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.380427] env[61964]: INFO nova.compute.claims [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.386966] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Build of instance 894ad0a3-851a-4b5e-82b0-050cbbf264f6 was re-scheduled: Binding failed for port 98482e59-e602-4e29-9a8a-ab8c13d042a6, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 561.386966] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 561.386966] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquiring lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.386966] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Acquired lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.387324] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.669989] env[61964]: INFO nova.compute.manager [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] [instance: 6317e489-9347-46b7-989a-f543af500b39] Took 1.03 seconds to deallocate network for instance. [ 561.922955] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.031037] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.533794] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Releasing lock "refresh_cache-894ad0a3-851a-4b5e-82b0-050cbbf264f6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.534043] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 562.534179] env[61964]: DEBUG nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 562.534361] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.568967] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.678395] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquiring lock "fd4fe83c-a976-4dd8-a8a0-24f8545292e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.678691] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "fd4fe83c-a976-4dd8-a8a0-24f8545292e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.714396] env[61964]: INFO nova.scheduler.client.report [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Deleted allocations for instance 6317e489-9347-46b7-989a-f543af500b39 [ 562.742433] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5540948f-8130-417b-902f-e0136f93148c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.753488] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa10d2e2-952c-47c3-9f6f-ec1b3e93d0ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.789728] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cec4e83-32d1-4c32-86cf-fc3d90a43cf8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.797564] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d42ef8-ef4c-4ad1-983a-0e74dd9cfa8c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.813485] env[61964]: DEBUG nova.compute.provider_tree [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.073507] env[61964]: DEBUG nova.network.neutron [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.148902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "4e14a519-b543-4a18-9e1d-7df16abf5345" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.149186] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.152042] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "4e14a519-b543-4a18-9e1d-7df16abf5345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.152042] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.152042] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.152705] env[61964]: INFO nova.compute.manager [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Terminating instance [ 563.206045] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "aa526175-3381-48e8-bd5c-73ed37931dbd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.206045] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "aa526175-3381-48e8-bd5c-73ed37931dbd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.225531] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05d23ce4-a551-446e-9fdd-9cb632235b75 tempest-FloatingIPsAssociationNegativeTestJSON-519276348 tempest-FloatingIPsAssociationNegativeTestJSON-519276348-project-member] Lock "6317e489-9347-46b7-989a-f543af500b39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.518s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.317151] env[61964]: DEBUG nova.scheduler.client.report [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 563.578758] env[61964]: INFO nova.compute.manager [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] [instance: 894ad0a3-851a-4b5e-82b0-050cbbf264f6] Took 1.04 seconds to deallocate network for instance. [ 563.658968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "refresh_cache-4e14a519-b543-4a18-9e1d-7df16abf5345" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.659345] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquired lock "refresh_cache-4e14a519-b543-4a18-9e1d-7df16abf5345" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.659621] env[61964]: DEBUG nova.network.neutron [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 563.729673] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 563.823507] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.823873] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 563.826824] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.067s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.143459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "10a51687-3a7c-4293-9710-c043e754cd25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.143801] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "10a51687-3a7c-4293-9710-c043e754cd25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.198809] env[61964]: DEBUG nova.network.neutron [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.262677] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.334966] env[61964]: DEBUG nova.compute.utils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.341766] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 564.341941] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.367390] env[61964]: DEBUG nova.network.neutron [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.464714] env[61964]: DEBUG nova.policy [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dff7fc465c16475dbe8bc35b93741700', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb22af0400f143ad920b0a844c2e5b76', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.619946] env[61964]: INFO nova.scheduler.client.report [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Deleted allocations for instance 894ad0a3-851a-4b5e-82b0-050cbbf264f6 [ 564.701159] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1893f284-4ab2-4b39-ab84-9ddf8e3fefa3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.710644] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd178548-7732-4cdc-a8e0-6b751401503d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.748969] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2393f322-35f9-478a-8f89-492d48d6ed77 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.759070] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8377d8e2-61b3-438f-a53c-39f4af38ccc2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.772587] env[61964]: DEBUG nova.compute.provider_tree [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.842886] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 564.869650] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Releasing lock "refresh_cache-4e14a519-b543-4a18-9e1d-7df16abf5345" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.873351] env[61964]: DEBUG nova.compute.manager [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 564.874872] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 564.874872] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2036a90-aff9-437f-8807-3abd17d17e43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.885926] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 564.886217] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1ded5c3-2f5a-4ff8-a7ca-e67cb52a9886 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.896548] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 564.896548] env[61964]: value = "task-1040531" [ 564.896548] env[61964]: _type = "Task" [ 564.896548] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.909437] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.134018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c95295ae-46ec-48f2-adf7-48ec813e58c0 tempest-TenantUsagesTestJSON-558391164 tempest-TenantUsagesTestJSON-558391164-project-member] Lock "894ad0a3-851a-4b5e-82b0-050cbbf264f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.006s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.155728] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Successfully created port: 0dbeceb4-f892-4cbe-ba33-4ba91117d88c {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.278244] env[61964]: DEBUG nova.scheduler.client.report [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 565.410127] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040531, 'name': PowerOffVM_Task, 'duration_secs': 0.139167} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.412208] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 565.412995] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 565.414370] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c7ad112-918a-4cff-bfe3-9309bafd26d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.439807] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 565.439807] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 565.439807] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleting the datastore file [datastore1] 4e14a519-b543-4a18-9e1d-7df16abf5345 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 565.439807] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed07af9e-276b-4932-b275-f4f70d89a6e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.447909] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for the task: (returnval){ [ 565.447909] env[61964]: value = "task-1040533" [ 565.447909] env[61964]: _type = "Task" [ 565.447909] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.458526] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.639361] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 565.785891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.786561] env[61964]: ERROR nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Traceback (most recent call last): [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.driver.spawn(context, instance, image_meta, [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] vm_ref = self.build_virtual_machine(instance, [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.786561] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] for vif in network_info: [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self._sync_wrapper(fn, *args, **kwargs) [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.wait() [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self[:] = self._gt.wait() [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self._exit_event.wait() [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] result = hub.switch() [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.787065] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return self.greenlet.switch() [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] result = function(*args, **kwargs) [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] return func(*args, **kwargs) [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise e [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] nwinfo = self.network_api.allocate_for_instance( [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] created_port_ids = self._update_ports_for_instance( [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] with excutils.save_and_reraise_exception(): [ 565.787865] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] self.force_reraise() [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise self.value [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] updated_port = self._update_port( [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] _ensure_no_port_binding_failure(port) [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] raise exception.PortBindingFailed(port_id=port['id']) [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] nova.exception.PortBindingFailed: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. [ 565.788343] env[61964]: ERROR nova.compute.manager [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] [ 565.788735] env[61964]: DEBUG nova.compute.utils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.793531] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Build of instance 51b42c8a-f780-4725-85ab-3fad0d621dfd was re-scheduled: Binding failed for port f4628cb6-2328-403d-aa29-6dc2c4191ef1, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.793743] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.793994] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquiring lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.794159] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Acquired lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.794329] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.797554] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.798736] env[61964]: INFO nova.compute.claims [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.802998] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "3a26e916-3d1e-4831-8351-0c12fc775398" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.803282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "3a26e916-3d1e-4831-8351-0c12fc775398" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.855173] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 565.878491] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 565.878740] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.878967] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 565.879204] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.879292] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 565.879470] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 565.879572] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 565.879775] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 565.879978] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 565.880208] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 565.880307] env[61964]: DEBUG nova.virt.hardware [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 565.881101] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841f1119-fe99-4bdd-9718-e4142d14ce63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.891441] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b144da0c-f1dc-4f2f-a5b4-9d5033af8a12 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.958376] env[61964]: DEBUG oslo_vmware.api [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Task: {'id': task-1040533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096521} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.958646] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 565.959117] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 565.959375] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 565.959571] env[61964]: INFO nova.compute.manager [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Took 1.09 seconds to destroy the instance on the hypervisor. [ 565.959829] env[61964]: DEBUG oslo.service.loopingcall [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.960068] env[61964]: DEBUG nova.compute.manager [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 565.960164] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.981809] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.176713] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.327254] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.485336] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.596018] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.988570] env[61964]: INFO nova.compute.manager [-] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Took 1.03 seconds to deallocate network for instance. [ 567.100867] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Releasing lock "refresh_cache-51b42c8a-f780-4725-85ab-3fad0d621dfd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.102302] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 567.103539] env[61964]: DEBUG nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 567.103539] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.148544] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.154373] env[61964]: ERROR nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 567.154373] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.154373] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.154373] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.154373] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.154373] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.154373] env[61964]: ERROR nova.compute.manager raise self.value [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.154373] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 567.154373] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.154373] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 567.154843] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.154843] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 567.154843] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 567.154843] env[61964]: ERROR nova.compute.manager [ 567.155963] env[61964]: Traceback (most recent call last): [ 567.155963] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 567.155963] env[61964]: listener.cb(fileno) [ 567.155963] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.155963] env[61964]: result = function(*args, **kwargs) [ 567.155963] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.155963] env[61964]: return func(*args, **kwargs) [ 567.155963] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 567.155963] env[61964]: raise e [ 567.155963] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.155963] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 567.155963] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.155963] env[61964]: created_port_ids = self._update_ports_for_instance( [ 567.155963] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.155963] env[61964]: with excutils.save_and_reraise_exception(): [ 567.155963] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.155963] env[61964]: self.force_reraise() [ 567.155963] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.155963] env[61964]: raise self.value [ 567.155963] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.155963] env[61964]: updated_port = self._update_port( [ 567.155963] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.155963] env[61964]: _ensure_no_port_binding_failure(port) [ 567.155963] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.155963] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 567.155963] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 567.155963] env[61964]: Removing descriptor: 17 [ 567.156922] env[61964]: ERROR nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Traceback (most recent call last): [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] yield resources [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.driver.spawn(context, instance, image_meta, [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] vm_ref = self.build_virtual_machine(instance, [ 567.156922] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] for vif in network_info: [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self._sync_wrapper(fn, *args, **kwargs) [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.wait() [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self[:] = self._gt.wait() [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self._exit_event.wait() [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.157410] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] result = hub.switch() [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self.greenlet.switch() [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] result = function(*args, **kwargs) [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return func(*args, **kwargs) [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise e [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] nwinfo = self.network_api.allocate_for_instance( [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] created_port_ids = self._update_ports_for_instance( [ 567.157808] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] with excutils.save_and_reraise_exception(): [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.force_reraise() [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise self.value [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] updated_port = self._update_port( [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] _ensure_no_port_binding_failure(port) [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise exception.PortBindingFailed(port_id=port['id']) [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 567.158229] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] [ 567.158640] env[61964]: INFO nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Terminating instance [ 567.168036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bb40c7-9553-4c0b-8242-d38788bd223f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.177969] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1ac464-c3ad-4b56-adeb-c38bf7895fc5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.213104] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bbc87f-7b75-43e5-ab25-f8a3515a7d9b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.223586] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c117c5b4-e952-47e2-83ec-3ea1b66e1fd3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.239719] env[61964]: DEBUG nova.compute.provider_tree [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.420435] env[61964]: DEBUG nova.compute.manager [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Received event network-changed-0dbeceb4-f892-4cbe-ba33-4ba91117d88c {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 567.420687] env[61964]: DEBUG nova.compute.manager [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Refreshing instance network info cache due to event network-changed-0dbeceb4-f892-4cbe-ba33-4ba91117d88c. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 567.421401] env[61964]: DEBUG oslo_concurrency.lockutils [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] Acquiring lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.421649] env[61964]: DEBUG oslo_concurrency.lockutils [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] Acquired lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.422411] env[61964]: DEBUG nova.network.neutron [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Refreshing network info cache for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 567.495831] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.656190] env[61964]: DEBUG nova.network.neutron [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.664966] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquiring lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.743185] env[61964]: DEBUG nova.scheduler.client.report [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 567.960769] env[61964]: DEBUG nova.network.neutron [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.155884] env[61964]: DEBUG nova.network.neutron [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.161747] env[61964]: INFO nova.compute.manager [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] [instance: 51b42c8a-f780-4725-85ab-3fad0d621dfd] Took 1.06 seconds to deallocate network for instance. [ 568.247528] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.248174] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 568.250769] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.255s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.662666] env[61964]: DEBUG oslo_concurrency.lockutils [req-12f63418-b59a-4375-9f18-7b01695a3220 req-ae726fa6-a54c-4780-b58c-c709f0bc59da service nova] Releasing lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.663080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquired lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.663273] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.761738] env[61964]: DEBUG nova.compute.utils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.771024] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 568.771024] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 568.933143] env[61964]: DEBUG nova.policy [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbcb6948440e47c1b88eb0002d6931ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '486fd4a100d14be6839a137f480069e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.139135] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f8f723-ab7a-40fc-adb6-da51cad39ba2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.149548] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb89d0b-31fb-47d6-87dc-094f2b555d8a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.199541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f12dde-ef0e-4ad0-89b8-787fbd275232 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.209416] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993ae1a3-74f2-453a-addc-906f0be131ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.227766] env[61964]: DEBUG nova.compute.provider_tree [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.229111] env[61964]: INFO nova.scheduler.client.report [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Deleted allocations for instance 51b42c8a-f780-4725-85ab-3fad0d621dfd [ 569.253526] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.269659] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 569.639658] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.737582] env[61964]: DEBUG nova.scheduler.client.report [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 569.741009] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d0759e17-e513-400c-b732-310c1c949683 tempest-ServerActionsTestOtherA-431855893 tempest-ServerActionsTestOtherA-431855893-project-member] Lock "51b42c8a-f780-4725-85ab-3fad0d621dfd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.485s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.146019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Releasing lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.146019] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 570.146019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 570.146019] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08d6059d-8bd0-4198-a135-5ba0d3e02a5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.153352] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20feb21c-e9d5-44d8-a266-74c3dea8ab91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.178246] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5 could not be found. [ 570.178666] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 570.179297] env[61964]: INFO nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 570.180037] env[61964]: DEBUG oslo.service.loopingcall [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.180037] env[61964]: DEBUG nova.compute.manager [-] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 570.180037] env[61964]: DEBUG nova.network.neutron [-] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.243999] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.244756] env[61964]: ERROR nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Traceback (most recent call last): [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.driver.spawn(context, instance, image_meta, [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] vm_ref = self.build_virtual_machine(instance, [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.244756] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] for vif in network_info: [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self._sync_wrapper(fn, *args, **kwargs) [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.wait() [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self[:] = self._gt.wait() [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self._exit_event.wait() [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] result = hub.switch() [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.245398] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return self.greenlet.switch() [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] result = function(*args, **kwargs) [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] return func(*args, **kwargs) [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise e [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] nwinfo = self.network_api.allocate_for_instance( [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] created_port_ids = self._update_ports_for_instance( [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] with excutils.save_and_reraise_exception(): [ 570.246089] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] self.force_reraise() [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise self.value [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] updated_port = self._update_port( [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] _ensure_no_port_binding_failure(port) [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] raise exception.PortBindingFailed(port_id=port['id']) [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] nova.exception.PortBindingFailed: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. [ 570.246695] env[61964]: ERROR nova.compute.manager [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] [ 570.247256] env[61964]: DEBUG nova.compute.utils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.247398] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 570.253234] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Build of instance ee058557-cace-4e32-bdf9-bd4af7fe29ff was re-scheduled: Binding failed for port 5109dac7-31e2-4a20-be1a-aa626808b400, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 570.255616] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 570.255616] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.255616] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.255616] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.256293] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.717s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.256492] env[61964]: DEBUG nova.objects.instance [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 570.279488] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 570.289509] env[61964]: DEBUG nova.network.neutron [-] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.318081] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 570.318458] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.318711] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 570.318951] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.319184] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 570.319503] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 570.319784] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 570.320147] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 570.320400] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 570.320643] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 570.320884] env[61964]: DEBUG nova.virt.hardware [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 570.322034] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3ad9c1-f5ac-4cda-831c-b1a4d4f84413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.332319] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d1b7d2-b9e4-492c-bf30-0e8eebbdefd2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.350183] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Successfully created port: 47138044-8990-415b-903a-079e38e0de53 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.375195] env[61964]: DEBUG nova.compute.manager [req-02de5be4-a356-4326-8ebb-109a3bdde37c req-46f5a65a-d81c-45f4-b841-e027525b7ad2 service nova] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Received event network-vif-deleted-0dbeceb4-f892-4cbe-ba33-4ba91117d88c {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 570.792472] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.793975] env[61964]: DEBUG nova.network.neutron [-] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.821178] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.043403] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.282978] env[61964]: DEBUG oslo_concurrency.lockutils [None req-52075112-a024-4a1d-a427-a271832315ab tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.282978] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.354s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.288316] env[61964]: INFO nova.compute.claims [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 571.299335] env[61964]: INFO nova.compute.manager [-] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Took 1.12 seconds to deallocate network for instance. [ 571.302750] env[61964]: DEBUG nova.compute.claims [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.302940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.549481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-ee058557-cace-4e32-bdf9-bd4af7fe29ff" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.549726] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 571.549890] env[61964]: DEBUG nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 571.550266] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.595565] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.098415] env[61964]: DEBUG nova.network.neutron [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.602609] env[61964]: INFO nova.compute.manager [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: ee058557-cace-4e32-bdf9-bd4af7fe29ff] Took 1.05 seconds to deallocate network for instance. [ 572.714228] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78dffe8-0240-48f7-aee0-2af31e9b8140 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.728797] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8f122d-9697-44f0-996d-171c8831779d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.763827] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43227a8e-0e6d-4220-9ff6-4898c70864c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.773119] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccce77fa-32f7-4132-aec3-8f24850dcc42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.788075] env[61964]: DEBUG nova.compute.provider_tree [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.295743] env[61964]: DEBUG nova.scheduler.client.report [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 573.655030] env[61964]: INFO nova.scheduler.client.report [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Deleted allocations for instance ee058557-cace-4e32-bdf9-bd4af7fe29ff [ 573.802815] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.802815] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 573.805745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.808s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.163627] env[61964]: DEBUG oslo_concurrency.lockutils [None req-364b129a-3cd7-48ff-acd3-3d39d905e51d tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "ee058557-cace-4e32-bdf9-bd4af7fe29ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.787s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.313873] env[61964]: DEBUG nova.compute.utils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 574.323639] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 574.323855] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 574.420779] env[61964]: ERROR nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 574.420779] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.420779] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.420779] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.420779] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.420779] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.420779] env[61964]: ERROR nova.compute.manager raise self.value [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.420779] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.420779] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.420779] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.421388] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.421388] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.421388] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 574.421388] env[61964]: ERROR nova.compute.manager [ 574.421388] env[61964]: Traceback (most recent call last): [ 574.421388] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.421388] env[61964]: listener.cb(fileno) [ 574.421388] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.421388] env[61964]: result = function(*args, **kwargs) [ 574.421388] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.421388] env[61964]: return func(*args, **kwargs) [ 574.421388] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 574.421388] env[61964]: raise e [ 574.421388] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.421388] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 574.421388] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.421388] env[61964]: created_port_ids = self._update_ports_for_instance( [ 574.421388] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.421388] env[61964]: with excutils.save_and_reraise_exception(): [ 574.421388] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.421388] env[61964]: self.force_reraise() [ 574.421388] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.421388] env[61964]: raise self.value [ 574.421388] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.421388] env[61964]: updated_port = self._update_port( [ 574.421388] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.421388] env[61964]: _ensure_no_port_binding_failure(port) [ 574.421388] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.421388] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.422254] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 574.422254] env[61964]: Removing descriptor: 17 [ 574.427690] env[61964]: ERROR nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Traceback (most recent call last): [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] yield resources [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.driver.spawn(context, instance, image_meta, [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] vm_ref = self.build_virtual_machine(instance, [ 574.427690] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] for vif in network_info: [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self._sync_wrapper(fn, *args, **kwargs) [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.wait() [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self[:] = self._gt.wait() [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self._exit_event.wait() [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.428272] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] result = hub.switch() [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self.greenlet.switch() [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] result = function(*args, **kwargs) [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return func(*args, **kwargs) [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise e [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] nwinfo = self.network_api.allocate_for_instance( [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] created_port_ids = self._update_ports_for_instance( [ 574.428759] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] with excutils.save_and_reraise_exception(): [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.force_reraise() [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise self.value [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] updated_port = self._update_port( [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] _ensure_no_port_binding_failure(port) [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise exception.PortBindingFailed(port_id=port['id']) [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 574.429185] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] [ 574.429582] env[61964]: INFO nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Terminating instance [ 574.517935] env[61964]: DEBUG nova.policy [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8582989ebff541ecb53cc9e66dd2e238', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b27cc5d81fa4adb8a1415bd33fa94c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 574.595282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.595670] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.677045] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 574.720427] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d828db46-46f6-464b-9a5f-a2dbd958377d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.737637] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475dc7d8-d1f7-4ecc-88ca-656f48022351 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.772323] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7bbcaa-891c-4dfd-8c2a-82cb6158ce50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.779912] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64645304-d2f1-470c-8da3-6256d5154ee5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.794273] env[61964]: DEBUG nova.compute.provider_tree [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.830805] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 574.938711] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.939514] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquired lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.939514] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.041029] env[61964]: DEBUG nova.compute.manager [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Received event network-changed-47138044-8990-415b-903a-079e38e0de53 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 575.041390] env[61964]: DEBUG nova.compute.manager [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Refreshing instance network info cache due to event network-changed-47138044-8990-415b-903a-079e38e0de53. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 575.041629] env[61964]: DEBUG oslo_concurrency.lockutils [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] Acquiring lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.213494] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.297408] env[61964]: DEBUG nova.scheduler.client.report [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 575.486428] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.678230] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.722110] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Successfully created port: 67ebdacc-29d5-4a65-a00b-ae2b584e7b89 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 575.805724] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.806601] env[61964]: ERROR nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Traceback (most recent call last): [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.driver.spawn(context, instance, image_meta, [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] vm_ref = self.build_virtual_machine(instance, [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.806601] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] for vif in network_info: [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self._sync_wrapper(fn, *args, **kwargs) [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.wait() [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self[:] = self._gt.wait() [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self._exit_event.wait() [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] result = hub.switch() [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.807016] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return self.greenlet.switch() [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] result = function(*args, **kwargs) [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] return func(*args, **kwargs) [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise e [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] nwinfo = self.network_api.allocate_for_instance( [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] created_port_ids = self._update_ports_for_instance( [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] with excutils.save_and_reraise_exception(): [ 575.807365] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] self.force_reraise() [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise self.value [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] updated_port = self._update_port( [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] _ensure_no_port_binding_failure(port) [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] raise exception.PortBindingFailed(port_id=port['id']) [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] nova.exception.PortBindingFailed: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. [ 575.807734] env[61964]: ERROR nova.compute.manager [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] [ 575.808050] env[61964]: DEBUG nova.compute.utils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.809175] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.832s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.810905] env[61964]: INFO nova.compute.claims [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.814953] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Build of instance c9e79aac-19af-4339-9b06-5907d2ac716c was re-scheduled: Binding failed for port 5a6b83b1-d922-4373-8923-2f297a695e94, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 575.815342] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 575.815510] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquiring lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.815667] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Acquired lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.815915] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.842767] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 575.874573] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 575.874830] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.874830] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 575.876522] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.876983] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 575.876983] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 575.877106] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 575.880576] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 575.880576] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 575.880576] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 575.880576] env[61964]: DEBUG nova.virt.hardware [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 575.880576] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91e2228-e72c-484b-934e-00311dc2c01f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.895734] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbf07a7-780a-4efb-94f6-b1df6c53794d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.183270] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Releasing lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.183835] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 576.184042] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 576.184401] env[61964]: DEBUG oslo_concurrency.lockutils [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] Acquired lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.184610] env[61964]: DEBUG nova.network.neutron [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Refreshing network info cache for port 47138044-8990-415b-903a-079e38e0de53 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 576.185977] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25ff6530-b249-422c-bfdd-5ddb593fb38b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.198361] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4600e7c-ada6-4fe9-a7ba-57ecceb35f02 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.223526] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c93818ca-3f7c-4cd2-b8c4-d7714c4404a2 could not be found. [ 576.223756] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 576.224016] env[61964]: INFO nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 576.224296] env[61964]: DEBUG oslo.service.loopingcall [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.224517] env[61964]: DEBUG nova.compute.manager [-] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 576.224626] env[61964]: DEBUG nova.network.neutron [-] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.261026] env[61964]: DEBUG nova.network.neutron [-] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.370319] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.627124] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.722837] env[61964]: DEBUG nova.network.neutron [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.766537] env[61964]: DEBUG nova.network.neutron [-] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.924180] env[61964]: DEBUG nova.network.neutron [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.132574] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Releasing lock "refresh_cache-c9e79aac-19af-4339-9b06-5907d2ac716c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.132574] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.132574] env[61964]: DEBUG nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.132574] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.157643] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "9f216f01-dc9f-459f-88be-3a5439d3701a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.157643] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.206759] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.215120] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b123b278-88c8-48cd-9e47-3f7b70efb4d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.225697] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a638205-ae81-4877-a16c-1dfc33a87a35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.262418] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d245665f-f141-47a0-a417-4aaa2ff62f2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.270538] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26eb0bb5-d8fe-49a5-b140-1b77bea0c41c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.278367] env[61964]: INFO nova.compute.manager [-] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Took 1.05 seconds to deallocate network for instance. [ 577.278602] env[61964]: DEBUG nova.compute.claims [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.278905] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.290599] env[61964]: DEBUG nova.compute.provider_tree [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.427313] env[61964]: DEBUG oslo_concurrency.lockutils [req-605abf34-967c-4996-bd28-34cc345dd3d8 req-f0ddf196-9672-4297-a525-cd915e35984a service nova] Releasing lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.717916] env[61964]: DEBUG nova.network.neutron [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.793884] env[61964]: DEBUG nova.scheduler.client.report [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 577.903804] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "0f1ed823-1531-4505-a68a-cb172f804655" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.904450] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.223442] env[61964]: INFO nova.compute.manager [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] [instance: c9e79aac-19af-4339-9b06-5907d2ac716c] Took 1.09 seconds to deallocate network for instance. [ 578.300628] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.304144] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.283s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.375038] env[61964]: DEBUG nova.compute.manager [req-ceaa8141-3ff9-40d7-afcb-2b2e295635c1 req-cda0906f-68bb-4980-8cd9-88b967a46f4c service nova] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Received event network-vif-deleted-47138044-8990-415b-903a-079e38e0de53 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 578.439061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "06631c59-9969-4763-bfe8-472e8dcf8848" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.439061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.504764] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.505093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.709515] env[61964]: ERROR nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 578.709515] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.709515] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.709515] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.709515] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.709515] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.709515] env[61964]: ERROR nova.compute.manager raise self.value [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.709515] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.709515] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.709515] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.710023] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.710023] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.710023] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 578.710023] env[61964]: ERROR nova.compute.manager [ 578.710023] env[61964]: Traceback (most recent call last): [ 578.710023] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.710023] env[61964]: listener.cb(fileno) [ 578.710023] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.710023] env[61964]: result = function(*args, **kwargs) [ 578.710023] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.710023] env[61964]: return func(*args, **kwargs) [ 578.710023] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 578.710023] env[61964]: raise e [ 578.710023] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.710023] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 578.710023] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.710023] env[61964]: created_port_ids = self._update_ports_for_instance( [ 578.710023] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.710023] env[61964]: with excutils.save_and_reraise_exception(): [ 578.710023] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.710023] env[61964]: self.force_reraise() [ 578.710023] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.710023] env[61964]: raise self.value [ 578.710023] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.710023] env[61964]: updated_port = self._update_port( [ 578.710023] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.710023] env[61964]: _ensure_no_port_binding_failure(port) [ 578.710023] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.710023] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.710823] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 578.710823] env[61964]: Removing descriptor: 15 [ 578.710823] env[61964]: ERROR nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Traceback (most recent call last): [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] yield resources [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.driver.spawn(context, instance, image_meta, [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.710823] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] vm_ref = self.build_virtual_machine(instance, [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] for vif in network_info: [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self._sync_wrapper(fn, *args, **kwargs) [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.wait() [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self[:] = self._gt.wait() [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self._exit_event.wait() [ 578.711218] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] result = hub.switch() [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self.greenlet.switch() [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] result = function(*args, **kwargs) [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return func(*args, **kwargs) [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise e [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] nwinfo = self.network_api.allocate_for_instance( [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.711677] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] created_port_ids = self._update_ports_for_instance( [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] with excutils.save_and_reraise_exception(): [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.force_reraise() [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise self.value [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] updated_port = self._update_port( [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] _ensure_no_port_binding_failure(port) [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.712560] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise exception.PortBindingFailed(port_id=port['id']) [ 578.712976] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 578.712976] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] [ 578.712976] env[61964]: INFO nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Terminating instance [ 578.809472] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "989b79d1-e3fa-40a7-9910-22ff1af8b657" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.809832] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "989b79d1-e3fa-40a7-9910-22ff1af8b657" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.221072] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquiring lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.221225] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquired lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.221404] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.230211] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477a7f11-d5eb-4738-8587-83c9a82c16a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.243185] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cfcf07-59b3-4632-9a3d-868892382140 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.288988] env[61964]: INFO nova.scheduler.client.report [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Deleted allocations for instance c9e79aac-19af-4339-9b06-5907d2ac716c [ 579.297306] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a38cc2-b1ce-456a-8883-a9ea48379abc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.303911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc939291-d9cb-4187-9aa9-a351241a98c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.318875] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "989b79d1-e3fa-40a7-9910-22ff1af8b657" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.509s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.319453] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 579.322321] env[61964]: DEBUG nova.compute.provider_tree [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.760271] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.799619] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dcd86cb2-1426-463c-bb0c-889d2950e50a tempest-InstanceActionsV221TestJSON-584289556 tempest-InstanceActionsV221TestJSON-584289556-project-member] Lock "c9e79aac-19af-4339-9b06-5907d2ac716c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.150s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.826169] env[61964]: DEBUG nova.scheduler.client.report [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 579.832748] env[61964]: DEBUG nova.compute.utils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.833348] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 579.833443] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 579.952906] env[61964]: DEBUG nova.policy [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '533387fa5c2745a29c3b9e1a22a4cb64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0cc7cf71f35f4e2eae669c623eb2c1b7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.958487] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.305728] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 580.333394] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 580.340022] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.340022] env[61964]: ERROR nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Traceback (most recent call last): [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.driver.spawn(context, instance, image_meta, [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.340022] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] vm_ref = self.build_virtual_machine(instance, [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] for vif in network_info: [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self._sync_wrapper(fn, *args, **kwargs) [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.wait() [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self[:] = self._gt.wait() [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self._exit_event.wait() [ 580.340390] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] result = hub.switch() [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return self.greenlet.switch() [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] result = function(*args, **kwargs) [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] return func(*args, **kwargs) [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise e [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] nwinfo = self.network_api.allocate_for_instance( [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.340746] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] created_port_ids = self._update_ports_for_instance( [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] with excutils.save_and_reraise_exception(): [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] self.force_reraise() [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise self.value [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] updated_port = self._update_port( [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] _ensure_no_port_binding_failure(port) [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.341143] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] raise exception.PortBindingFailed(port_id=port['id']) [ 580.341499] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] nova.exception.PortBindingFailed: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. [ 580.341499] env[61964]: ERROR nova.compute.manager [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] [ 580.341499] env[61964]: DEBUG nova.compute.utils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.342062] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Build of instance 07771b38-3dc9-4608-8bea-48971bbc9651 was re-scheduled: Binding failed for port 24baf232-a5bb-4017-a72b-2545bf292cce, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 580.343026] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 580.343026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.346027] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquired lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.346027] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.346027] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.248s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.346027] env[61964]: DEBUG nova.objects.instance [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 580.461839] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Releasing lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.463088] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 580.463088] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.463088] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccbab692-8306-44d6-9a49-9e299eb59ea5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.475924] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db9ca5c-e4c5-4c76-9ed1-3f016019eabd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.502650] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 84cda217-8acd-4fd9-80ac-cb83edcc1347 could not be found. [ 580.502896] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.503092] env[61964]: INFO nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.503363] env[61964]: DEBUG oslo.service.loopingcall [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.504756] env[61964]: DEBUG nova.compute.manager [-] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 580.504861] env[61964]: DEBUG nova.network.neutron [-] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.510860] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.510860] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.552965] env[61964]: DEBUG nova.network.neutron [-] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.847795] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.902025] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.056376] env[61964]: DEBUG nova.network.neutron [-] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.133015] env[61964]: DEBUG nova.compute.manager [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Received event network-changed-67ebdacc-29d5-4a65-a00b-ae2b584e7b89 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 581.133206] env[61964]: DEBUG nova.compute.manager [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Refreshing instance network info cache due to event network-changed-67ebdacc-29d5-4a65-a00b-ae2b584e7b89. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 581.133635] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] Acquiring lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.133635] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] Acquired lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.133813] env[61964]: DEBUG nova.network.neutron [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Refreshing network info cache for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 581.164775] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Successfully created port: 96342de6-1e83-47c8-8f49-bc213fac27f7 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.207081] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.355604] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 581.360572] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7c5f3508-0711-4ef4-90e5-15c593bea9aa tempest-ServersAdmin275Test-214770473 tempest-ServersAdmin275Test-214770473-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.361956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.100s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.363646] env[61964]: INFO nova.compute.claims [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.389764] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 581.389996] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.390175] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 581.390370] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.391165] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 581.391165] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 581.391376] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 581.392080] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 581.392080] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 581.392080] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 581.392080] env[61964]: DEBUG nova.virt.hardware [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 581.395591] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62642d9-67da-495e-9d93-d9e5a74c0967 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.404063] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256fc853-ec3e-44bd-82c9-528067e8a073 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.561976] env[61964]: INFO nova.compute.manager [-] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Took 1.06 seconds to deallocate network for instance. [ 581.563926] env[61964]: DEBUG nova.compute.claims [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.564240] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.689059] env[61964]: DEBUG nova.network.neutron [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.709662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Releasing lock "refresh_cache-07771b38-3dc9-4608-8bea-48971bbc9651" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.709866] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 581.710072] env[61964]: DEBUG nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 581.710246] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.733964] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.158103] env[61964]: DEBUG nova.network.neutron [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.188442] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "65566c20-6d69-471c-b098-3c30c01d9955" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.189034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.236649] env[61964]: DEBUG nova.network.neutron [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.665317] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] Releasing lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.665317] env[61964]: DEBUG nova.compute.manager [req-9c0da016-c3a4-4234-9b9e-aa1b80b71847 req-322dbd58-ba0d-4f61-9572-4fce406d56e0 service nova] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Received event network-vif-deleted-67ebdacc-29d5-4a65-a00b-ae2b584e7b89 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 582.744396] env[61964]: INFO nova.compute.manager [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: 07771b38-3dc9-4608-8bea-48971bbc9651] Took 1.03 seconds to deallocate network for instance. [ 582.814173] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8484d997-af4e-4fce-bf98-d8ed2d085557 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.822098] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4174028-0228-467c-9c08-ef0a7871f126 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.861030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ad6f02-be81-4af3-a172-39851de397d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.868551] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6849d72-f35d-45fc-91f1-f817523cbf20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.884140] env[61964]: DEBUG nova.compute.provider_tree [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.388561] env[61964]: DEBUG nova.scheduler.client.report [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 583.782133] env[61964]: INFO nova.scheduler.client.report [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Deleted allocations for instance 07771b38-3dc9-4608-8bea-48971bbc9651 [ 583.895498] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.895498] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 583.899866] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.723s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.902229] env[61964]: INFO nova.compute.claims [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.294675] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e6b19a0c-b267-4f8e-a651-6615575c15c1 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "07771b38-3dc9-4608-8bea-48971bbc9651" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.755s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.409109] env[61964]: DEBUG nova.compute.utils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.415911] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 584.415911] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 584.543205] env[61964]: DEBUG nova.policy [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7126b60ca87c4818a115d74a2dfe52a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78a4982fafd64fe6ae728311969c511d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.608404] env[61964]: DEBUG nova.compute.manager [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Received event network-changed-96342de6-1e83-47c8-8f49-bc213fac27f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 584.608597] env[61964]: DEBUG nova.compute.manager [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Refreshing instance network info cache due to event network-changed-96342de6-1e83-47c8-8f49-bc213fac27f7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 584.610247] env[61964]: DEBUG oslo_concurrency.lockutils [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] Acquiring lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.610647] env[61964]: DEBUG oslo_concurrency.lockutils [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] Acquired lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.610849] env[61964]: DEBUG nova.network.neutron [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Refreshing network info cache for port 96342de6-1e83-47c8-8f49-bc213fac27f7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 584.800745] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 584.920156] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 585.011192] env[61964]: ERROR nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 585.011192] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.011192] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.011192] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.011192] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.011192] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.011192] env[61964]: ERROR nova.compute.manager raise self.value [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.011192] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.011192] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.011192] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.011734] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.011734] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.011734] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 585.011734] env[61964]: ERROR nova.compute.manager [ 585.011734] env[61964]: Traceback (most recent call last): [ 585.011734] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.011734] env[61964]: listener.cb(fileno) [ 585.011734] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.011734] env[61964]: result = function(*args, **kwargs) [ 585.011734] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.011734] env[61964]: return func(*args, **kwargs) [ 585.011734] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.011734] env[61964]: raise e [ 585.011734] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.011734] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 585.011734] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.011734] env[61964]: created_port_ids = self._update_ports_for_instance( [ 585.011734] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.011734] env[61964]: with excutils.save_and_reraise_exception(): [ 585.011734] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.011734] env[61964]: self.force_reraise() [ 585.011734] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.011734] env[61964]: raise self.value [ 585.011734] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.011734] env[61964]: updated_port = self._update_port( [ 585.011734] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.011734] env[61964]: _ensure_no_port_binding_failure(port) [ 585.011734] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.011734] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.012671] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 585.012671] env[61964]: Removing descriptor: 17 [ 585.012671] env[61964]: ERROR nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Traceback (most recent call last): [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] yield resources [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.driver.spawn(context, instance, image_meta, [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.012671] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] vm_ref = self.build_virtual_machine(instance, [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] for vif in network_info: [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self._sync_wrapper(fn, *args, **kwargs) [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.wait() [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self[:] = self._gt.wait() [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self._exit_event.wait() [ 585.013073] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] result = hub.switch() [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self.greenlet.switch() [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] result = function(*args, **kwargs) [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return func(*args, **kwargs) [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise e [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] nwinfo = self.network_api.allocate_for_instance( [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.013452] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] created_port_ids = self._update_ports_for_instance( [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] with excutils.save_and_reraise_exception(): [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.force_reraise() [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise self.value [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] updated_port = self._update_port( [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] _ensure_no_port_binding_failure(port) [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.013801] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise exception.PortBindingFailed(port_id=port['id']) [ 585.014171] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 585.014171] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] [ 585.014171] env[61964]: INFO nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Terminating instance [ 585.199687] env[61964]: DEBUG nova.network.neutron [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.320617] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.362071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c0d937-334f-442d-a095-c0bcb20ce8ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.373065] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb04566-fc98-47b0-8ca8-85c5e0b76a5d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.403924] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3395682-594c-408e-9b30-f727aa03ce69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.410396] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a94dff-b3cf-4a45-b759-67f3b0a17334 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.428345] env[61964]: DEBUG nova.compute.provider_tree [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.520553] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.643734] env[61964]: DEBUG nova.network.neutron [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.929881] env[61964]: DEBUG nova.scheduler.client.report [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 585.942716] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 585.958394] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Successfully created port: 13c13bbf-8177-4595-b94c-42f1fade3ecd {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.977270] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:20:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='853702721',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1424805346',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 585.977270] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.977270] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 585.977457] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.977457] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 585.977732] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 585.977732] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 585.977868] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 585.979841] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 585.980135] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 585.980348] env[61964]: DEBUG nova.virt.hardware [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 585.981587] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be19b56d-411e-4e07-b942-7dd82c7f8136 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.990811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c5f52a-1bef-43a0-a931-512f2b00a3c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.146909] env[61964]: DEBUG oslo_concurrency.lockutils [req-ede1dcb8-4818-4358-81d2-265b3c4dc862 req-b926639e-611a-4776-ba9e-1f969179c676 service nova] Releasing lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.147364] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquired lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.147483] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.387844] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.387844] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.440202] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.443275] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 586.448743] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.950s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.449082] env[61964]: DEBUG nova.objects.instance [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lazy-loading 'resources' on Instance uuid 4e14a519-b543-4a18-9e1d-7df16abf5345 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 586.595095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.595241] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.634569] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.634951] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.733948] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.954180] env[61964]: DEBUG nova.compute.utils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.956152] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 586.956327] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 587.048546] env[61964]: DEBUG nova.policy [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57332e5de0924a21ac302ee901b2455e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd2bf2040b9e4119859d869347d57426', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.148314] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.420311] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e75e365-03a0-4e81-aa55-3f1eb4d5faaa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.425152] env[61964]: DEBUG nova.compute.manager [req-701cec55-96b1-4d15-9f10-0eb5d58dfb5f req-65d247c4-7982-418f-b4a0-e9f9a2900775 service nova] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Received event network-vif-deleted-96342de6-1e83-47c8-8f49-bc213fac27f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 587.432026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2113fbf2-ccfd-4614-bbbf-6b06ef85ecb7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.468360] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 587.477023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3950b305-229f-41f9-8eaa-5f12f23679b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.479073] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4d5fcc-f5e8-46df-add5-b1ef0672bd6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.497395] env[61964]: DEBUG nova.compute.provider_tree [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.609376] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Successfully created port: b12ce13d-f9c5-4120-b28f-7decdcb17c31 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.655829] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Releasing lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.656467] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 587.656467] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.656682] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7991348b-7ee6-4fd9-9af8-843e2ac39b71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.665916] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9410602-8f0b-44bd-85e3-fa5138590029 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.695175] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e could not be found. [ 587.695423] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.695611] env[61964]: INFO nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.695861] env[61964]: DEBUG oslo.service.loopingcall [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.696184] env[61964]: DEBUG nova.compute.manager [-] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 587.696262] env[61964]: DEBUG nova.network.neutron [-] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.909500] env[61964]: DEBUG nova.network.neutron [-] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.000764] env[61964]: DEBUG nova.scheduler.client.report [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 588.148681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.148898] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.415718] env[61964]: DEBUG nova.network.neutron [-] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.483338] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 588.510723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.063s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.513156] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.721s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.514837] env[61964]: INFO nova.compute.claims [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.529112] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 588.529225] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.529307] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 588.529471] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.529610] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 588.529751] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 588.529943] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 588.530784] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 588.531412] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 588.531589] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 588.531761] env[61964]: DEBUG nova.virt.hardware [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 588.532671] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98852b02-975d-4ca2-a871-6862dff4ad7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.541230] env[61964]: INFO nova.scheduler.client.report [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Deleted allocations for instance 4e14a519-b543-4a18-9e1d-7df16abf5345 [ 588.550215] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2806b041-17c9-43cb-b742-9d02de1d25b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.810651] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "16276be9-b305-4d1d-afde-bc98be42687a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.810900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.919342] env[61964]: INFO nova.compute.manager [-] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Took 1.22 seconds to deallocate network for instance. [ 588.923396] env[61964]: DEBUG nova.compute.claims [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.923984] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.051512] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3553af01-ed73-45ac-96df-d89b2412cd28 tempest-ServersAdmin275Test-564113191 tempest-ServersAdmin275Test-564113191-project-member] Lock "4e14a519-b543-4a18-9e1d-7df16abf5345" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.902s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.568578] env[61964]: DEBUG nova.compute.manager [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Received event network-changed-13c13bbf-8177-4595-b94c-42f1fade3ecd {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 589.570095] env[61964]: DEBUG nova.compute.manager [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Refreshing instance network info cache due to event network-changed-13c13bbf-8177-4595-b94c-42f1fade3ecd. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 589.570095] env[61964]: DEBUG oslo_concurrency.lockutils [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] Acquiring lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.570095] env[61964]: DEBUG oslo_concurrency.lockutils [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] Acquired lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.570095] env[61964]: DEBUG nova.network.neutron [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Refreshing network info cache for port 13c13bbf-8177-4595-b94c-42f1fade3ecd {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 589.620782] env[61964]: ERROR nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 589.620782] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.620782] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.620782] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.620782] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.620782] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.620782] env[61964]: ERROR nova.compute.manager raise self.value [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.620782] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.620782] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.620782] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.621300] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.621300] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.621300] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 589.621300] env[61964]: ERROR nova.compute.manager [ 589.621300] env[61964]: Traceback (most recent call last): [ 589.621300] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.621300] env[61964]: listener.cb(fileno) [ 589.621300] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.621300] env[61964]: result = function(*args, **kwargs) [ 589.621300] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.621300] env[61964]: return func(*args, **kwargs) [ 589.621300] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.621300] env[61964]: raise e [ 589.621300] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.621300] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 589.621300] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.621300] env[61964]: created_port_ids = self._update_ports_for_instance( [ 589.621300] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.621300] env[61964]: with excutils.save_and_reraise_exception(): [ 589.621300] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.621300] env[61964]: self.force_reraise() [ 589.621300] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.621300] env[61964]: raise self.value [ 589.621300] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.621300] env[61964]: updated_port = self._update_port( [ 589.621300] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.621300] env[61964]: _ensure_no_port_binding_failure(port) [ 589.621300] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.621300] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.622091] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 589.622091] env[61964]: Removing descriptor: 15 [ 589.622091] env[61964]: ERROR nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Traceback (most recent call last): [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] yield resources [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.driver.spawn(context, instance, image_meta, [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.622091] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] vm_ref = self.build_virtual_machine(instance, [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] for vif in network_info: [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self._sync_wrapper(fn, *args, **kwargs) [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.wait() [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self[:] = self._gt.wait() [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self._exit_event.wait() [ 589.622524] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] result = hub.switch() [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self.greenlet.switch() [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] result = function(*args, **kwargs) [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return func(*args, **kwargs) [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise e [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] nwinfo = self.network_api.allocate_for_instance( [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.622885] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] created_port_ids = self._update_ports_for_instance( [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] with excutils.save_and_reraise_exception(): [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.force_reraise() [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise self.value [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] updated_port = self._update_port( [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] _ensure_no_port_binding_failure(port) [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.623271] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise exception.PortBindingFailed(port_id=port['id']) [ 589.623661] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 589.623661] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] [ 589.623661] env[61964]: INFO nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Terminating instance [ 590.070727] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8eb160-0a8f-4bb7-833a-632b60a428d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.097528] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22295877-2033-4d57-bceb-e8bd8e3b1b10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.131958] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquiring lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.134122] env[61964]: DEBUG nova.network.neutron [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.136517] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d8fef6-4bcf-4edd-a19d-f6fba400f46f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.141957] env[61964]: DEBUG nova.compute.manager [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Received event network-changed-b12ce13d-f9c5-4120-b28f-7decdcb17c31 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 590.141957] env[61964]: DEBUG nova.compute.manager [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Refreshing instance network info cache due to event network-changed-b12ce13d-f9c5-4120-b28f-7decdcb17c31. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 590.141957] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] Acquiring lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.141957] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] Acquired lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.141957] env[61964]: DEBUG nova.network.neutron [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Refreshing network info cache for port b12ce13d-f9c5-4120-b28f-7decdcb17c31 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 590.150447] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4b548e-3c24-42fe-af29-0e4610334f16 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.167461] env[61964]: DEBUG nova.compute.provider_tree [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.258710] env[61964]: DEBUG nova.network.neutron [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.578379] env[61964]: ERROR nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 590.578379] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 590.578379] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.578379] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.578379] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.578379] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.578379] env[61964]: ERROR nova.compute.manager raise self.value [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.578379] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.578379] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.578379] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.579260] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.579260] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.579260] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 590.579260] env[61964]: ERROR nova.compute.manager [ 590.579260] env[61964]: Traceback (most recent call last): [ 590.579260] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.579260] env[61964]: listener.cb(fileno) [ 590.579260] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.579260] env[61964]: result = function(*args, **kwargs) [ 590.579260] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.579260] env[61964]: return func(*args, **kwargs) [ 590.579260] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 590.579260] env[61964]: raise e [ 590.579260] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 590.579260] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 590.579260] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.579260] env[61964]: created_port_ids = self._update_ports_for_instance( [ 590.579260] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.579260] env[61964]: with excutils.save_and_reraise_exception(): [ 590.579260] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.579260] env[61964]: self.force_reraise() [ 590.579260] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.579260] env[61964]: raise self.value [ 590.579260] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.579260] env[61964]: updated_port = self._update_port( [ 590.579260] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.579260] env[61964]: _ensure_no_port_binding_failure(port) [ 590.579260] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.579260] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.580262] env[61964]: nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 590.580262] env[61964]: Removing descriptor: 18 [ 590.580262] env[61964]: ERROR nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Traceback (most recent call last): [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] yield resources [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.driver.spawn(context, instance, image_meta, [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.580262] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] vm_ref = self.build_virtual_machine(instance, [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] for vif in network_info: [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self._sync_wrapper(fn, *args, **kwargs) [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.wait() [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self[:] = self._gt.wait() [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self._exit_event.wait() [ 590.580648] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] result = hub.switch() [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self.greenlet.switch() [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] result = function(*args, **kwargs) [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return func(*args, **kwargs) [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise e [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] nwinfo = self.network_api.allocate_for_instance( [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.581045] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] created_port_ids = self._update_ports_for_instance( [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] with excutils.save_and_reraise_exception(): [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.force_reraise() [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise self.value [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] updated_port = self._update_port( [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] _ensure_no_port_binding_failure(port) [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.581441] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise exception.PortBindingFailed(port_id=port['id']) [ 590.581806] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 590.581806] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] [ 590.581806] env[61964]: INFO nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Terminating instance [ 590.674170] env[61964]: DEBUG nova.scheduler.client.report [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 590.697407] env[61964]: DEBUG nova.network.neutron [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.762786] env[61964]: DEBUG oslo_concurrency.lockutils [req-67a1131c-e7f5-47b2-a454-e9e52804cbd3 req-fd2023d8-30c1-44f8-8355-663e882a5b9b service nova] Releasing lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.763270] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquired lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.763630] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.879392] env[61964]: DEBUG nova.network.neutron [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.086682] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquiring lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.181772] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.668s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.181933] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 591.185879] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.882s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.302453] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.381460] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c4f9cb6-0a18-4d6d-9664-e47a5e0380ec req-57851c61-c4ce-461c-a180-974db43b759e service nova] Releasing lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.381753] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquired lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.381944] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.473991] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.690596] env[61964]: DEBUG nova.compute.utils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.692565] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 591.692868] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 591.770308] env[61964]: DEBUG nova.policy [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96091947177e4d488c9a467703fbbba4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c884f8125a3148158e04a8f5589f0381', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.777838] env[61964]: DEBUG nova.compute.manager [req-73bf5301-5cf0-4ae3-afc4-cd4e6361c670 req-d12b8903-7a73-4849-b557-05a651cd59f7 service nova] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Received event network-vif-deleted-13c13bbf-8177-4595-b94c-42f1fade3ecd {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 591.871786] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "a2a25787-1426-4c09-892d-395e1becccc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.872035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.911511] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.977896] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Releasing lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.978405] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 591.978602] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 591.978894] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8f39301-4121-41f1-91e2-56ea9fabdc62 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.989164] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.996745] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6283731-86ea-43f3-8163-08e02fea4ee8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.009469] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Releasing lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.010135] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 592.010307] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 592.013953] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d288d6f7-7a45-4226-9d00-302d212a737e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.024736] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4864ae-adb0-4d7f-9332-42a40f850186 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.045360] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 08bd2fb8-fc8d-440c-9369-196605a894e7 could not be found. [ 592.045583] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.045763] env[61964]: INFO nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Took 0.07 seconds to destroy the instance on the hypervisor. [ 592.046009] env[61964]: DEBUG oslo.service.loopingcall [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.046768] env[61964]: DEBUG nova.compute.manager [-] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 592.046848] env[61964]: DEBUG nova.network.neutron [-] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.059328] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e95ab551-63c1-4fa2-b615-5c6a543c39c1 could not be found. [ 592.059515] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.059693] env[61964]: INFO nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 592.059920] env[61964]: DEBUG oslo.service.loopingcall [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.062750] env[61964]: DEBUG nova.compute.manager [-] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 592.062750] env[61964]: DEBUG nova.network.neutron [-] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.065271] env[61964]: DEBUG nova.network.neutron [-] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.086137] env[61964]: DEBUG nova.network.neutron [-] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.166805] env[61964]: DEBUG nova.compute.manager [req-3f7f4216-3da2-4dd8-91e4-8eb545cd4886 req-0ae26ebb-291d-46e8-8264-ac7da9dc7836 service nova] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Received event network-vif-deleted-b12ce13d-f9c5-4120-b28f-7decdcb17c31 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 592.193542] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 592.239199] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a779a9df-20b8-4b9a-b444-18fb564de449 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.248669] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa39a8e-3e77-4917-9aed-e96d7ca20637 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.285082] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c836bfba-68eb-46ed-85de-5ec77ab9543f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.299019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e4d664-d83e-4cae-be8d-6d039a126aa1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.312786] env[61964]: DEBUG nova.compute.provider_tree [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.318781] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Successfully created port: f7398661-c69f-42c9-b154-a0710632d443 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.568345] env[61964]: DEBUG nova.network.neutron [-] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.586954] env[61964]: DEBUG nova.network.neutron [-] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.816877] env[61964]: DEBUG nova.scheduler.client.report [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 593.071752] env[61964]: INFO nova.compute.manager [-] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Took 1.02 seconds to deallocate network for instance. [ 593.076336] env[61964]: DEBUG nova.compute.claims [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.076336] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.089558] env[61964]: INFO nova.compute.manager [-] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Took 1.03 seconds to deallocate network for instance. [ 593.091859] env[61964]: DEBUG nova.compute.claims [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.091980] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.212020] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 593.248045] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 593.248392] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.248598] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 593.249045] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.249045] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 593.249157] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 593.249457] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 593.249622] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 593.249827] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 593.250978] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 593.250978] env[61964]: DEBUG nova.virt.hardware [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 593.251821] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3ec7ef-32cb-4598-995d-548e2f0c4c99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.264304] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1375548e-99cb-4eaa-b106-8ba1c2811178 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.322909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.138s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.325408] env[61964]: ERROR nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Traceback (most recent call last): [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.driver.spawn(context, instance, image_meta, [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] vm_ref = self.build_virtual_machine(instance, [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.325408] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] for vif in network_info: [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self._sync_wrapper(fn, *args, **kwargs) [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.wait() [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self[:] = self._gt.wait() [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self._exit_event.wait() [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] result = hub.switch() [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.325994] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return self.greenlet.switch() [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] result = function(*args, **kwargs) [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] return func(*args, **kwargs) [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise e [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] nwinfo = self.network_api.allocate_for_instance( [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] created_port_ids = self._update_ports_for_instance( [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] with excutils.save_and_reraise_exception(): [ 593.326427] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] self.force_reraise() [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise self.value [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] updated_port = self._update_port( [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] _ensure_no_port_binding_failure(port) [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] raise exception.PortBindingFailed(port_id=port['id']) [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] nova.exception.PortBindingFailed: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. [ 593.326805] env[61964]: ERROR nova.compute.manager [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] [ 593.327137] env[61964]: DEBUG nova.compute.utils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.331538] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.115s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.331538] env[61964]: INFO nova.compute.claims [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.342447] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Build of instance bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5 was re-scheduled: Binding failed for port 0dbeceb4-f892-4cbe-ba33-4ba91117d88c, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 593.342447] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 593.342447] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquiring lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.342447] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Acquired lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.342985] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.874283] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.007500] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.512777] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Releasing lock "refresh_cache-bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.512901] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 594.513238] env[61964]: DEBUG nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 594.513467] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.540936] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.549068] env[61964]: ERROR nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 594.549068] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.549068] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.549068] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.549068] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.549068] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.549068] env[61964]: ERROR nova.compute.manager raise self.value [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.549068] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.549068] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.549068] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.549735] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.549735] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.549735] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 594.549735] env[61964]: ERROR nova.compute.manager [ 594.549735] env[61964]: Traceback (most recent call last): [ 594.549735] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.549735] env[61964]: listener.cb(fileno) [ 594.549735] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.549735] env[61964]: result = function(*args, **kwargs) [ 594.549735] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.549735] env[61964]: return func(*args, **kwargs) [ 594.549735] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.549735] env[61964]: raise e [ 594.549735] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.549735] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 594.549735] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.549735] env[61964]: created_port_ids = self._update_ports_for_instance( [ 594.549735] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.549735] env[61964]: with excutils.save_and_reraise_exception(): [ 594.549735] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.549735] env[61964]: self.force_reraise() [ 594.549735] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.549735] env[61964]: raise self.value [ 594.549735] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.549735] env[61964]: updated_port = self._update_port( [ 594.549735] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.549735] env[61964]: _ensure_no_port_binding_failure(port) [ 594.549735] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.549735] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.552313] env[61964]: nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 594.552313] env[61964]: Removing descriptor: 18 [ 594.552313] env[61964]: ERROR nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Traceback (most recent call last): [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] yield resources [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.driver.spawn(context, instance, image_meta, [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.552313] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] vm_ref = self.build_virtual_machine(instance, [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] for vif in network_info: [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self._sync_wrapper(fn, *args, **kwargs) [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.wait() [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self[:] = self._gt.wait() [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self._exit_event.wait() [ 594.552963] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] result = hub.switch() [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self.greenlet.switch() [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] result = function(*args, **kwargs) [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return func(*args, **kwargs) [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise e [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] nwinfo = self.network_api.allocate_for_instance( [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.553580] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] created_port_ids = self._update_ports_for_instance( [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] with excutils.save_and_reraise_exception(): [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.force_reraise() [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise self.value [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] updated_port = self._update_port( [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] _ensure_no_port_binding_failure(port) [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.554017] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise exception.PortBindingFailed(port_id=port['id']) [ 594.554399] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 594.554399] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] [ 594.554399] env[61964]: INFO nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Terminating instance [ 594.616608] env[61964]: DEBUG nova.compute.manager [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Received event network-changed-f7398661-c69f-42c9-b154-a0710632d443 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 594.616608] env[61964]: DEBUG nova.compute.manager [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Refreshing instance network info cache due to event network-changed-f7398661-c69f-42c9-b154-a0710632d443. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 594.616812] env[61964]: DEBUG oslo_concurrency.lockutils [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] Acquiring lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.616958] env[61964]: DEBUG oslo_concurrency.lockutils [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] Acquired lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.617177] env[61964]: DEBUG nova.network.neutron [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Refreshing network info cache for port f7398661-c69f-42c9-b154-a0710632d443 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 594.724944] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.725259] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.845136] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14ad1bd-bb0d-40ad-ab1d-a5eb9422511d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.854244] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd0611c-f364-4206-b6b9-78d7c556c3bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.888018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e25a7fa-55c7-41db-bf35-0ec5bef5d5e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.894290] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d556d83a-4f66-470f-bce7-58d6a268c791 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.907432] env[61964]: DEBUG nova.compute.provider_tree [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.045887] env[61964]: DEBUG nova.network.neutron [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.058556] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquiring lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.142036] env[61964]: DEBUG nova.network.neutron [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.391416] env[61964]: DEBUG nova.network.neutron [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.409936] env[61964]: DEBUG nova.scheduler.client.report [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 595.547467] env[61964]: INFO nova.compute.manager [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] [instance: bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5] Took 1.03 seconds to deallocate network for instance. [ 595.894399] env[61964]: DEBUG oslo_concurrency.lockutils [req-306a8c24-15bd-4b0d-a2e0-2bea2be7b4b5 req-d05e6548-7ce9-47fe-9e0b-f3c847d014b9 service nova] Releasing lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.894907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquired lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.895055] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.921394] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.921969] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 595.925565] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.647s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.432184] env[61964]: DEBUG nova.compute.utils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.436707] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 596.440319] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.471402] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.519903] env[61964]: DEBUG nova.policy [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e47cf22875b4d72b48f978eb1b8fc7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab4edd9ffe804eca8b8b8a9aeac4bb41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.580662] env[61964]: INFO nova.scheduler.client.report [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Deleted allocations for instance bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5 [ 596.604173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.604173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.864708] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.925926] env[61964]: DEBUG nova.compute.manager [req-fc66bbc5-e28f-40f8-a351-9b6390517e7c req-bd290442-b418-4880-ac6d-c1e61a20c0b0 service nova] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Received event network-vif-deleted-f7398661-c69f-42c9-b154-a0710632d443 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 596.940501] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 596.980660] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be921594-915b-4a07-ade0-0852cd1801b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.993258] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3723736a-f407-4050-a68b-bb69e5813b44 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.033770] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8063e05-94ec-4860-95bf-9874f3984436 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.038334] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Successfully created port: 7e708664-b5e3-4342-aebf-067a4037a464 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.044717] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28beb52b-a2ba-47ad-afff-1f2029237308 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.059165] env[61964]: DEBUG nova.compute.provider_tree [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.093212] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8519c80a-d623-4c4b-8ffa-4769e49b1813 tempest-ImagesOneServerTestJSON-94103413 tempest-ImagesOneServerTestJSON-94103413-project-member] Lock "bfb0c91d-da34-4edd-b2be-94dc5b5dd4a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.406s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.366547] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Releasing lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.366653] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 597.366841] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 597.367186] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33190c81-169d-4bfb-89be-a30ecf804a42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.377933] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b977b3-6a8f-4489-8dc1-312dd63fc4ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.403576] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 19a6958e-4887-459b-8bf7-dd4cbbedb700 could not be found. [ 597.403774] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 597.403959] env[61964]: INFO nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Took 0.04 seconds to destroy the instance on the hypervisor. [ 597.404226] env[61964]: DEBUG oslo.service.loopingcall [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.404466] env[61964]: DEBUG nova.compute.manager [-] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.404582] env[61964]: DEBUG nova.network.neutron [-] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 597.421453] env[61964]: DEBUG nova.network.neutron [-] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.566474] env[61964]: DEBUG nova.scheduler.client.report [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.594721] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 597.927458] env[61964]: DEBUG nova.network.neutron [-] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.952908] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 597.990044] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 597.990346] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.990506] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 597.990913] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.990913] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 597.991012] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 597.991270] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 597.991481] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 597.995016] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 597.995016] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 597.995016] env[61964]: DEBUG nova.virt.hardware [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 597.995016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6140d550-4a06-4bed-b295-1590d5356f60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.001803] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c660e70c-ce01-492c-92d9-d4d64365b727 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.070994] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.145s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.071685] env[61964]: ERROR nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Traceback (most recent call last): [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.driver.spawn(context, instance, image_meta, [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] vm_ref = self.build_virtual_machine(instance, [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.071685] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] for vif in network_info: [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self._sync_wrapper(fn, *args, **kwargs) [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.wait() [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self[:] = self._gt.wait() [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self._exit_event.wait() [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] result = hub.switch() [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.071963] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return self.greenlet.switch() [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] result = function(*args, **kwargs) [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] return func(*args, **kwargs) [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise e [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] nwinfo = self.network_api.allocate_for_instance( [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] created_port_ids = self._update_ports_for_instance( [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] with excutils.save_and_reraise_exception(): [ 598.072321] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] self.force_reraise() [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise self.value [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] updated_port = self._update_port( [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] _ensure_no_port_binding_failure(port) [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] raise exception.PortBindingFailed(port_id=port['id']) [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] nova.exception.PortBindingFailed: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. [ 598.072638] env[61964]: ERROR nova.compute.manager [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] [ 598.073325] env[61964]: DEBUG nova.compute.utils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.075726] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.228s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.077132] env[61964]: INFO nova.compute.claims [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.079883] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Build of instance c93818ca-3f7c-4cd2-b8c4-d7714c4404a2 was re-scheduled: Binding failed for port 47138044-8990-415b-903a-079e38e0de53, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 598.080375] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 598.081587] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquiring lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.081785] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Acquired lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.081955] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.124712] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.335834] env[61964]: DEBUG nova.compute.manager [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Received event network-changed-7e708664-b5e3-4342-aebf-067a4037a464 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 598.336065] env[61964]: DEBUG nova.compute.manager [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Refreshing instance network info cache due to event network-changed-7e708664-b5e3-4342-aebf-067a4037a464. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 598.336274] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] Acquiring lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.337219] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] Acquired lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.337759] env[61964]: DEBUG nova.network.neutron [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Refreshing network info cache for port 7e708664-b5e3-4342-aebf-067a4037a464 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 598.432397] env[61964]: INFO nova.compute.manager [-] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Took 1.03 seconds to deallocate network for instance. [ 598.434994] env[61964]: DEBUG nova.compute.claims [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.435424] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.609734] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.674497] env[61964]: ERROR nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 598.674497] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 598.674497] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.674497] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.674497] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.674497] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.674497] env[61964]: ERROR nova.compute.manager raise self.value [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.674497] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.674497] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.674497] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.675031] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.675031] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.675031] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 598.675031] env[61964]: ERROR nova.compute.manager [ 598.675031] env[61964]: Traceback (most recent call last): [ 598.675031] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.675031] env[61964]: listener.cb(fileno) [ 598.675031] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.675031] env[61964]: result = function(*args, **kwargs) [ 598.675031] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.675031] env[61964]: return func(*args, **kwargs) [ 598.675031] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 598.675031] env[61964]: raise e [ 598.675031] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 598.675031] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 598.675031] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.675031] env[61964]: created_port_ids = self._update_ports_for_instance( [ 598.675031] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.675031] env[61964]: with excutils.save_and_reraise_exception(): [ 598.675031] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.675031] env[61964]: self.force_reraise() [ 598.675031] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.675031] env[61964]: raise self.value [ 598.675031] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.675031] env[61964]: updated_port = self._update_port( [ 598.675031] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.675031] env[61964]: _ensure_no_port_binding_failure(port) [ 598.675031] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.675031] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.675669] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 598.675669] env[61964]: Removing descriptor: 17 [ 598.675669] env[61964]: ERROR nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Traceback (most recent call last): [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] yield resources [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.driver.spawn(context, instance, image_meta, [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.675669] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] vm_ref = self.build_virtual_machine(instance, [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] for vif in network_info: [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self._sync_wrapper(fn, *args, **kwargs) [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.wait() [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self[:] = self._gt.wait() [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self._exit_event.wait() [ 598.676194] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] result = hub.switch() [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self.greenlet.switch() [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] result = function(*args, **kwargs) [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return func(*args, **kwargs) [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise e [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] nwinfo = self.network_api.allocate_for_instance( [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.676753] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] created_port_ids = self._update_ports_for_instance( [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] with excutils.save_and_reraise_exception(): [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.force_reraise() [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise self.value [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] updated_port = self._update_port( [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] _ensure_no_port_binding_failure(port) [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.677257] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise exception.PortBindingFailed(port_id=port['id']) [ 598.677543] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 598.677543] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] [ 598.677543] env[61964]: INFO nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Terminating instance [ 598.707847] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.859141] env[61964]: DEBUG nova.network.neutron [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.923721] env[61964]: DEBUG nova.network.neutron [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.181833] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.214141] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Releasing lock "refresh_cache-c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.214141] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 599.214280] env[61964]: DEBUG nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 599.214466] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.235308] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.427608] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba37e19e-3374-45ea-9b2f-2882e2f95300 req-dfeeda2a-c747-4a47-818b-517fb9bf3fd7 service nova] Releasing lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.427608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquired lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.427608] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.581067] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d4e2bb-3a10-4a56-8f0d-e4a6a7e67b79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.587879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa18824-6c86-46c4-8dc2-3df4c7fa02a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.623382] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d2afdc-9501-45d1-a3cd-be0e371b82be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.631390] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b4fc6e-18b3-4256-a42d-45d7b98a8792 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.643839] env[61964]: DEBUG nova.compute.provider_tree [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.739583] env[61964]: DEBUG nova.network.neutron [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.953029] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.015046] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.150209] env[61964]: DEBUG nova.scheduler.client.report [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 600.244419] env[61964]: INFO nova.compute.manager [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] [instance: c93818ca-3f7c-4cd2-b8c4-d7714c4404a2] Took 1.03 seconds to deallocate network for instance. [ 600.363051] env[61964]: DEBUG nova.compute.manager [req-7e520bb7-85bc-4d8c-905a-dbb2ff207d5f req-06ac70cb-49e3-4cf5-8eef-fa52dca11c0b service nova] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Received event network-vif-deleted-7e708664-b5e3-4342-aebf-067a4037a464 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 600.521114] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Releasing lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.521567] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 600.521753] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 600.522079] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab2218d9-90f7-4293-9eaf-d1eca5e638f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.530750] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed5a5dd-a7ea-4a6e-bad7-aba6237148a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.552762] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c45bf0b-17c2-4065-94d9-52ee7b9af151 could not be found. [ 600.553278] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 600.553278] env[61964]: INFO nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Took 0.03 seconds to destroy the instance on the hypervisor. [ 600.553430] env[61964]: DEBUG oslo.service.loopingcall [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.553632] env[61964]: DEBUG nova.compute.manager [-] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 600.553725] env[61964]: DEBUG nova.network.neutron [-] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.577475] env[61964]: DEBUG nova.network.neutron [-] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.656638] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.581s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.657580] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 600.660394] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.096s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.083441] env[61964]: DEBUG nova.network.neutron [-] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.167798] env[61964]: DEBUG nova.compute.utils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.169382] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 601.170287] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 601.276092] env[61964]: DEBUG nova.policy [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42d7ff1aa0594b1591368d6a67c759a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1000445744e43c2972fa969dce31bd6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.277416] env[61964]: INFO nova.scheduler.client.report [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Deleted allocations for instance c93818ca-3f7c-4cd2-b8c4-d7714c4404a2 [ 601.588349] env[61964]: INFO nova.compute.manager [-] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Took 1.03 seconds to deallocate network for instance. [ 601.591215] env[61964]: DEBUG nova.compute.claims [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 601.591428] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.645233] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c6a682-9816-421b-9508-29fdfaa9e0a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.656588] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf07d7b1-d2a7-4a89-bb8f-2e195c5e7208 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.695762] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 601.700043] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93159f46-657e-482e-ad1e-f87c4d198f1b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.711218] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b2fd6c-6b05-4b38-af46-5de0c3716a95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.729937] env[61964]: DEBUG nova.compute.provider_tree [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.789112] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b60ada39-3763-4918-b8dc-b07f5c7e49b8 tempest-ServersAdminTestJSON-1794823353 tempest-ServersAdminTestJSON-1794823353-project-member] Lock "c93818ca-3f7c-4cd2-b8c4-d7714c4404a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.182s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.942635] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Successfully created port: adf548f1-d6d1-4418-97b3-68733c9aa028 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.233936] env[61964]: DEBUG nova.scheduler.client.report [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 602.294648] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 602.705577] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 602.735537] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 602.735784] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.735941] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 602.736141] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.736289] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 602.736480] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 602.736639] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 602.736794] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 602.736958] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 602.737331] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 602.738017] env[61964]: DEBUG nova.virt.hardware [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 602.738540] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944fe26f-1040-4d09-9f7e-91475564fa66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.742211] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.742874] env[61964]: ERROR nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Traceback (most recent call last): [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.driver.spawn(context, instance, image_meta, [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] vm_ref = self.build_virtual_machine(instance, [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.742874] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] for vif in network_info: [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self._sync_wrapper(fn, *args, **kwargs) [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.wait() [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self[:] = self._gt.wait() [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self._exit_event.wait() [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] result = hub.switch() [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.743156] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return self.greenlet.switch() [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] result = function(*args, **kwargs) [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] return func(*args, **kwargs) [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise e [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] nwinfo = self.network_api.allocate_for_instance( [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] created_port_ids = self._update_ports_for_instance( [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] with excutils.save_and_reraise_exception(): [ 602.743459] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] self.force_reraise() [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise self.value [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] updated_port = self._update_port( [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] _ensure_no_port_binding_failure(port) [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] raise exception.PortBindingFailed(port_id=port['id']) [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] nova.exception.PortBindingFailed: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. [ 602.743745] env[61964]: ERROR nova.compute.manager [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] [ 602.744143] env[61964]: DEBUG nova.compute.utils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.745605] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.425s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.746820] env[61964]: INFO nova.compute.claims [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.752375] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Build of instance 84cda217-8acd-4fd9-80ac-cb83edcc1347 was re-scheduled: Binding failed for port 67ebdacc-29d5-4a65-a00b-ae2b584e7b89, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 602.752375] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 602.752375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquiring lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.752375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Acquired lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.752576] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.759513] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8000c70f-c2e5-4cb3-9547-1f374b4754a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.815699] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.275597] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.392754] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.435089] env[61964]: DEBUG nova.compute.manager [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Received event network-changed-adf548f1-d6d1-4418-97b3-68733c9aa028 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 603.435551] env[61964]: DEBUG nova.compute.manager [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Refreshing instance network info cache due to event network-changed-adf548f1-d6d1-4418-97b3-68733c9aa028. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 603.435796] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] Acquiring lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.435969] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] Acquired lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.436995] env[61964]: DEBUG nova.network.neutron [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Refreshing network info cache for port adf548f1-d6d1-4418-97b3-68733c9aa028 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 603.611392] env[61964]: ERROR nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 603.611392] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.611392] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.611392] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.611392] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.611392] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.611392] env[61964]: ERROR nova.compute.manager raise self.value [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.611392] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.611392] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.611392] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.614480] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.614480] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.614480] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 603.614480] env[61964]: ERROR nova.compute.manager [ 603.614480] env[61964]: Traceback (most recent call last): [ 603.614480] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.614480] env[61964]: listener.cb(fileno) [ 603.614480] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.614480] env[61964]: result = function(*args, **kwargs) [ 603.614480] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.614480] env[61964]: return func(*args, **kwargs) [ 603.614480] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.614480] env[61964]: raise e [ 603.614480] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.614480] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 603.614480] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.614480] env[61964]: created_port_ids = self._update_ports_for_instance( [ 603.614480] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.614480] env[61964]: with excutils.save_and_reraise_exception(): [ 603.614480] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.614480] env[61964]: self.force_reraise() [ 603.614480] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.614480] env[61964]: raise self.value [ 603.614480] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.614480] env[61964]: updated_port = self._update_port( [ 603.614480] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.614480] env[61964]: _ensure_no_port_binding_failure(port) [ 603.614480] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.614480] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.615173] env[61964]: nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 603.615173] env[61964]: Removing descriptor: 18 [ 603.615173] env[61964]: ERROR nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Traceback (most recent call last): [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] yield resources [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.driver.spawn(context, instance, image_meta, [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.615173] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] vm_ref = self.build_virtual_machine(instance, [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] for vif in network_info: [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self._sync_wrapper(fn, *args, **kwargs) [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.wait() [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self[:] = self._gt.wait() [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self._exit_event.wait() [ 603.615458] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] result = hub.switch() [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self.greenlet.switch() [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] result = function(*args, **kwargs) [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return func(*args, **kwargs) [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise e [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] nwinfo = self.network_api.allocate_for_instance( [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.615771] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] created_port_ids = self._update_ports_for_instance( [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] with excutils.save_and_reraise_exception(): [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.force_reraise() [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise self.value [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] updated_port = self._update_port( [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] _ensure_no_port_binding_failure(port) [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.616061] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise exception.PortBindingFailed(port_id=port['id']) [ 603.616304] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 603.616304] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] [ 603.616304] env[61964]: INFO nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Terminating instance [ 603.895826] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Releasing lock "refresh_cache-84cda217-8acd-4fd9-80ac-cb83edcc1347" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.896111] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 603.896294] env[61964]: DEBUG nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.896455] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.930396] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.972373] env[61964]: DEBUG nova.network.neutron [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.093889] env[61964]: DEBUG nova.network.neutron [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.120591] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquiring lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.230500] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37bac84-4a3c-4569-85b3-364f1172219a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.238902] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc457e9-e4fd-4263-9fef-3b31d1bc293b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.277363] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932bc486-9e2d-48e0-85a9-b10a3bdccde5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.283477] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.283701] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.289557] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db0577f-c6fa-4e3b-abef-f81236e2605b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.306129] env[61964]: DEBUG nova.compute.provider_tree [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.434199] env[61964]: DEBUG nova.network.neutron [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.596088] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbf3faa6-c855-40fc-bb86-8cf484420bb3 req-2da5ba38-d367-4ba5-8584-eb4c29b98bd5 service nova] Releasing lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.596517] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquired lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.596702] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.809765] env[61964]: DEBUG nova.scheduler.client.report [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 604.936723] env[61964]: INFO nova.compute.manager [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] [instance: 84cda217-8acd-4fd9-80ac-cb83edcc1347] Took 1.04 seconds to deallocate network for instance. [ 605.116215] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.192845] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.314745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.315305] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 605.321506] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.398s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.462431] env[61964]: DEBUG nova.compute.manager [req-2ece965a-2aeb-478d-9342-fe7fe008af72 req-68f62616-6571-432f-b632-ffa206428c95 service nova] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Received event network-vif-deleted-adf548f1-d6d1-4418-97b3-68733c9aa028 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 605.695482] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Releasing lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.695905] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 605.696110] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 605.696414] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdf7c008-4b9a-46a4-892e-0d5b158e1008 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.706137] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddefe514-c03f-48a7-8d64-73e0f369548d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.729758] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4 could not be found. [ 605.729974] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 605.730171] env[61964]: INFO nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 605.730413] env[61964]: DEBUG oslo.service.loopingcall [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.730619] env[61964]: DEBUG nova.compute.manager [-] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 605.730709] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.745724] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.827939] env[61964]: DEBUG nova.compute.utils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.829770] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 605.829940] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.915278] env[61964]: DEBUG nova.policy [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b20910460ad240aeb9b702fd4bd0d160', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c117df1b68db404cb7009e76ebeecc93', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.977384] env[61964]: INFO nova.scheduler.client.report [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Deleted allocations for instance 84cda217-8acd-4fd9-80ac-cb83edcc1347 [ 606.248566] env[61964]: DEBUG nova.network.neutron [-] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.330589] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 606.345258] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Successfully created port: 5d254e55-abb4-49fe-9f38-6661ebc50141 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 606.370886] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6f2469-7998-419e-92ab-933175c29fa1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.382096] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0a1f61-7404-4352-9ef8-dbe00f11c353 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.417301] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2d7466-47e6-4ea3-bc55-2942dccec67e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.424890] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7affd2be-ed90-493e-b3f2-b5aed0a5eea1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.439247] env[61964]: DEBUG nova.compute.provider_tree [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.486173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03b2e7b7-ec7c-4d69-ba1f-a53e641ab5e7 tempest-VolumesAssistedSnapshotsTest-1008125522 tempest-VolumesAssistedSnapshotsTest-1008125522-project-member] Lock "84cda217-8acd-4fd9-80ac-cb83edcc1347" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.409s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.755442] env[61964]: INFO nova.compute.manager [-] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Took 1.02 seconds to deallocate network for instance. [ 606.756375] env[61964]: DEBUG nova.compute.claims [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.756699] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.942536] env[61964]: DEBUG nova.scheduler.client.report [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 606.992987] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 607.339936] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 607.380362] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 607.380804] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 607.381207] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 607.381631] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 607.381890] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 607.382173] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 607.382489] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 607.382739] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 607.385018] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 607.385018] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 607.385018] env[61964]: DEBUG nova.virt.hardware [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 607.385018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7067df11-bd36-4efb-830d-d9f9594e41d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.393786] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e61e72-0a59-445a-99f0-1006b0854c96 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.449904] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.128s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.454216] env[61964]: ERROR nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Traceback (most recent call last): [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.driver.spawn(context, instance, image_meta, [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] vm_ref = self.build_virtual_machine(instance, [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.454216] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] for vif in network_info: [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self._sync_wrapper(fn, *args, **kwargs) [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.wait() [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self[:] = self._gt.wait() [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self._exit_event.wait() [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] result = hub.switch() [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.454495] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return self.greenlet.switch() [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] result = function(*args, **kwargs) [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] return func(*args, **kwargs) [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise e [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] nwinfo = self.network_api.allocate_for_instance( [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] created_port_ids = self._update_ports_for_instance( [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] with excutils.save_and_reraise_exception(): [ 607.454796] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] self.force_reraise() [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise self.value [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] updated_port = self._update_port( [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] _ensure_no_port_binding_failure(port) [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] raise exception.PortBindingFailed(port_id=port['id']) [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] nova.exception.PortBindingFailed: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. [ 607.455102] env[61964]: ERROR nova.compute.manager [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] [ 607.455387] env[61964]: DEBUG nova.compute.utils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.455387] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Build of instance e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e was re-scheduled: Binding failed for port 96342de6-1e83-47c8-8f49-bc213fac27f7, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 607.455387] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 607.455387] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquiring lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.455515] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Acquired lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.455515] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.455639] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.381s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.520976] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.539591] env[61964]: DEBUG nova.compute.manager [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Received event network-changed-5d254e55-abb4-49fe-9f38-6661ebc50141 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 607.539986] env[61964]: DEBUG nova.compute.manager [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Refreshing instance network info cache due to event network-changed-5d254e55-abb4-49fe-9f38-6661ebc50141. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 607.540270] env[61964]: DEBUG oslo_concurrency.lockutils [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] Acquiring lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.540441] env[61964]: DEBUG oslo_concurrency.lockutils [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] Acquired lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.540578] env[61964]: DEBUG nova.network.neutron [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Refreshing network info cache for port 5d254e55-abb4-49fe-9f38-6661ebc50141 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 607.559719] env[61964]: ERROR nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 607.559719] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.559719] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.559719] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.559719] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.559719] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.559719] env[61964]: ERROR nova.compute.manager raise self.value [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.559719] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.559719] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.559719] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.560312] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.560312] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.560312] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 607.560312] env[61964]: ERROR nova.compute.manager [ 607.560312] env[61964]: Traceback (most recent call last): [ 607.560312] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.560312] env[61964]: listener.cb(fileno) [ 607.560312] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.560312] env[61964]: result = function(*args, **kwargs) [ 607.560312] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.560312] env[61964]: return func(*args, **kwargs) [ 607.560312] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.560312] env[61964]: raise e [ 607.560312] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.560312] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 607.560312] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.560312] env[61964]: created_port_ids = self._update_ports_for_instance( [ 607.560312] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.560312] env[61964]: with excutils.save_and_reraise_exception(): [ 607.560312] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.560312] env[61964]: self.force_reraise() [ 607.560312] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.560312] env[61964]: raise self.value [ 607.560312] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.560312] env[61964]: updated_port = self._update_port( [ 607.560312] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.560312] env[61964]: _ensure_no_port_binding_failure(port) [ 607.560312] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.560312] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.561632] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 607.561632] env[61964]: Removing descriptor: 18 [ 607.561632] env[61964]: ERROR nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Traceback (most recent call last): [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] yield resources [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.driver.spawn(context, instance, image_meta, [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.561632] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] vm_ref = self.build_virtual_machine(instance, [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] for vif in network_info: [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self._sync_wrapper(fn, *args, **kwargs) [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.wait() [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self[:] = self._gt.wait() [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self._exit_event.wait() [ 607.562115] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] result = hub.switch() [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self.greenlet.switch() [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] result = function(*args, **kwargs) [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return func(*args, **kwargs) [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise e [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] nwinfo = self.network_api.allocate_for_instance( [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.562558] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] created_port_ids = self._update_ports_for_instance( [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] with excutils.save_and_reraise_exception(): [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.force_reraise() [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise self.value [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] updated_port = self._update_port( [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] _ensure_no_port_binding_failure(port) [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.563375] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise exception.PortBindingFailed(port_id=port['id']) [ 607.563796] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 607.563796] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] [ 607.563796] env[61964]: INFO nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Terminating instance [ 608.207825] env[61964]: DEBUG nova.network.neutron [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.210620] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.275104] env[61964]: DEBUG nova.network.neutron [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.402167] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.651583] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8be107-6365-4360-92e7-33e613603e24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.659240] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59cfecef-6b28-4ed4-b426-5887721b14bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.692675] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.693710] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08295499-71ad-4dd4-b364-967becfefba1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.701147] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17dd04a-5137-424f-9719-87b022cf923c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.715697] env[61964]: DEBUG nova.compute.provider_tree [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.778227] env[61964]: DEBUG oslo_concurrency.lockutils [req-583a7ae2-eb21-421d-a918-6439d4ab4098 req-aa0fd127-e6b4-4d47-b6c0-19f3d6623ce1 service nova] Releasing lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.779344] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquired lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.779344] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.903509] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Releasing lock "refresh_cache-e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.903811] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 608.904007] env[61964]: DEBUG nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 608.904179] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.923517] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.219545] env[61964]: DEBUG nova.scheduler.client.report [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 609.306152] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.397964] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.425793] env[61964]: DEBUG nova.network.neutron [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.590828] env[61964]: DEBUG nova.compute.manager [req-7d6acc8c-206e-4bed-b4f4-36f30766e45a req-4d17312e-1caa-4305-8893-197a2a07c458 service nova] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Received event network-vif-deleted-5d254e55-abb4-49fe-9f38-6661ebc50141 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 609.727997] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.271s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.727997] env[61964]: ERROR nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Traceback (most recent call last): [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.driver.spawn(context, instance, image_meta, [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.727997] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] vm_ref = self.build_virtual_machine(instance, [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] for vif in network_info: [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self._sync_wrapper(fn, *args, **kwargs) [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.wait() [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self[:] = self._gt.wait() [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self._exit_event.wait() [ 609.728411] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] result = hub.switch() [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return self.greenlet.switch() [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] result = function(*args, **kwargs) [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] return func(*args, **kwargs) [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise e [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] nwinfo = self.network_api.allocate_for_instance( [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.728689] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] created_port_ids = self._update_ports_for_instance( [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] with excutils.save_and_reraise_exception(): [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] self.force_reraise() [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise self.value [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] updated_port = self._update_port( [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] _ensure_no_port_binding_failure(port) [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.728960] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] raise exception.PortBindingFailed(port_id=port['id']) [ 609.729272] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] nova.exception.PortBindingFailed: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. [ 609.729272] env[61964]: ERROR nova.compute.manager [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] [ 609.729272] env[61964]: DEBUG nova.compute.utils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.734018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.641s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.740168] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Build of instance 08bd2fb8-fc8d-440c-9369-196605a894e7 was re-scheduled: Binding failed for port 13c13bbf-8177-4595-b94c-42f1fade3ecd, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 609.740168] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 609.740168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquiring lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.740168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Acquired lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.740431] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.900996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Releasing lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.902482] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 609.902697] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.903944] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91c11c91-fc9a-433b-a34e-2b188241d355 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.915657] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af9033c-232d-43af-b160-3a5fedf5e7b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.931302] env[61964]: INFO nova.compute.manager [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] [instance: e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e] Took 1.03 seconds to deallocate network for instance. [ 609.947768] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 669fe2d9-cb37-4757-a199-18877cc6149e could not be found. [ 609.948033] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 609.948224] env[61964]: INFO nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 609.948478] env[61964]: DEBUG oslo.service.loopingcall [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.948974] env[61964]: DEBUG nova.compute.manager [-] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 609.949272] env[61964]: DEBUG nova.network.neutron [-] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.968102] env[61964]: DEBUG nova.network.neutron [-] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.279413] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.386848] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.471025] env[61964]: DEBUG nova.network.neutron [-] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.674096] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd60233-724a-443a-85df-9c3fa77430ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.682853] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62db10ea-4e1c-4233-9eca-c98467b82238 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.713252] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fb02ba-a680-4d6f-a04c-3b06a1bf894a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.721534] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a639e6-ed72-44d8-9b27-66b72f5fc517 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.735462] env[61964]: DEBUG nova.compute.provider_tree [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.891660] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Releasing lock "refresh_cache-08bd2fb8-fc8d-440c-9369-196605a894e7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.891904] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 610.892106] env[61964]: DEBUG nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 610.892305] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.907416] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.977498] env[61964]: INFO nova.compute.manager [-] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Took 1.03 seconds to deallocate network for instance. [ 610.982029] env[61964]: DEBUG nova.compute.claims [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.982193] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.989805] env[61964]: INFO nova.scheduler.client.report [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Deleted allocations for instance e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e [ 611.238494] env[61964]: DEBUG nova.scheduler.client.report [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 611.410154] env[61964]: DEBUG nova.network.neutron [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.498321] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0c816508-26aa-4d42-8d9f-ccda694b7d85 tempest-ServerGroupTestJSON-1871190936 tempest-ServerGroupTestJSON-1871190936-project-member] Lock "e4bbdfec-2c8b-4d17-8a0e-0a13e6f3659e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.401s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.748753] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.751680] env[61964]: ERROR nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Traceback (most recent call last): [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.driver.spawn(context, instance, image_meta, [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] vm_ref = self.build_virtual_machine(instance, [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.751680] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] for vif in network_info: [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self._sync_wrapper(fn, *args, **kwargs) [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.wait() [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self[:] = self._gt.wait() [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self._exit_event.wait() [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] result = hub.switch() [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.752022] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return self.greenlet.switch() [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] result = function(*args, **kwargs) [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] return func(*args, **kwargs) [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise e [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] nwinfo = self.network_api.allocate_for_instance( [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] created_port_ids = self._update_ports_for_instance( [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] with excutils.save_and_reraise_exception(): [ 611.752350] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] self.force_reraise() [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise self.value [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] updated_port = self._update_port( [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] _ensure_no_port_binding_failure(port) [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] raise exception.PortBindingFailed(port_id=port['id']) [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] nova.exception.PortBindingFailed: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. [ 611.752633] env[61964]: ERROR nova.compute.manager [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] [ 611.752869] env[61964]: DEBUG nova.compute.utils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.752869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.629s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.756469] env[61964]: INFO nova.compute.claims [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.761107] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Build of instance e95ab551-63c1-4fa2-b615-5c6a543c39c1 was re-scheduled: Binding failed for port b12ce13d-f9c5-4120-b28f-7decdcb17c31, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 611.761632] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 611.761869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquiring lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.762043] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Acquired lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.762220] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.923025] env[61964]: INFO nova.compute.manager [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] [instance: 08bd2fb8-fc8d-440c-9369-196605a894e7] Took 1.03 seconds to deallocate network for instance. [ 612.003446] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 612.293071] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.388998] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.541682] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.893535] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Releasing lock "refresh_cache-e95ab551-63c1-4fa2-b615-5c6a543c39c1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.894852] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 612.894852] env[61964]: DEBUG nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 612.894852] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.919324] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.969945] env[61964]: INFO nova.scheduler.client.report [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Deleted allocations for instance 08bd2fb8-fc8d-440c-9369-196605a894e7 [ 613.251214] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a671a4-53c4-4d0c-afbc-1f5d0865d1c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.260275] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c7b954-4bb3-4aa8-8bb9-cb963e96acda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.296623] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7989146-a5eb-408c-a581-ce2df917f4dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.305388] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490367f0-786d-4fc7-b5cc-bc9b61625502 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.319721] env[61964]: DEBUG nova.compute.provider_tree [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.424298] env[61964]: DEBUG nova.network.neutron [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.490176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b3cbb596-65cb-4c8f-ac93-089678266ddc tempest-ServersWithSpecificFlavorTestJSON-1410782215 tempest-ServersWithSpecificFlavorTestJSON-1410782215-project-member] Lock "08bd2fb8-fc8d-440c-9369-196605a894e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.415s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.823708] env[61964]: DEBUG nova.scheduler.client.report [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 613.929711] env[61964]: INFO nova.compute.manager [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] [instance: e95ab551-63c1-4fa2-b615-5c6a543c39c1] Took 1.03 seconds to deallocate network for instance. [ 613.993508] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 614.331298] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.331850] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.338029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.900s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.527369] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.837915] env[61964]: DEBUG nova.compute.utils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.838644] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 614.838644] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 614.971606] env[61964]: INFO nova.scheduler.client.report [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Deleted allocations for instance e95ab551-63c1-4fa2-b615-5c6a543c39c1 [ 615.009559] env[61964]: DEBUG nova.policy [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e47cf22875b4d72b48f978eb1b8fc7e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab4edd9ffe804eca8b8b8a9aeac4bb41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.310018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.310018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.347432] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.390586] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a7da6c-6fbc-4535-ac0c-9a513d08a8f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.400307] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4aa7fc-84ea-4ee1-bc6d-af09abc905e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.432252] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f5a260-5329-4ab5-91dc-378a75edff62 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.439811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35148a1e-0547-4aa0-b198-b9320f190413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.458431] env[61964]: DEBUG nova.compute.provider_tree [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.491100] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfb74d9b-1f15-456e-8c87-e370c0e6badc tempest-ServerDiagnosticsNegativeTest-104397940 tempest-ServerDiagnosticsNegativeTest-104397940-project-member] Lock "e95ab551-63c1-4fa2-b615-5c6a543c39c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.600s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.493215] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Successfully created port: 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.960508] env[61964]: DEBUG nova.scheduler.client.report [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 615.996628] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 616.360219] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.398279] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 616.398630] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.398825] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 616.398914] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.399079] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 616.400885] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 616.400885] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 616.400885] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 616.400885] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 616.400885] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 616.401036] env[61964]: DEBUG nova.virt.hardware [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 616.402227] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da76c13-9a7c-4d3f-8eff-89a289e17a4f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.412222] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a33857-59c0-4245-a184-a1b32875ad20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.476714] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.139s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.476714] env[61964]: ERROR nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Traceback (most recent call last): [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.driver.spawn(context, instance, image_meta, [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.476714] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] vm_ref = self.build_virtual_machine(instance, [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] for vif in network_info: [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self._sync_wrapper(fn, *args, **kwargs) [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.wait() [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self[:] = self._gt.wait() [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self._exit_event.wait() [ 616.477076] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] result = hub.switch() [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return self.greenlet.switch() [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] result = function(*args, **kwargs) [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] return func(*args, **kwargs) [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise e [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] nwinfo = self.network_api.allocate_for_instance( [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.477428] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] created_port_ids = self._update_ports_for_instance( [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] with excutils.save_and_reraise_exception(): [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] self.force_reraise() [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise self.value [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] updated_port = self._update_port( [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] _ensure_no_port_binding_failure(port) [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.477704] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] raise exception.PortBindingFailed(port_id=port['id']) [ 616.477986] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] nova.exception.PortBindingFailed: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. [ 616.477986] env[61964]: ERROR nova.compute.manager [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] [ 616.477986] env[61964]: DEBUG nova.compute.utils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.477986] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.885s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.483819] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Build of instance 19a6958e-4887-459b-8bf7-dd4cbbedb700 was re-scheduled: Binding failed for port f7398661-c69f-42c9-b154-a0710632d443, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 616.484575] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 616.485402] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquiring lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.485402] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Acquired lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.485402] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.534463] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.175303] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.209910] env[61964]: ERROR nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 617.209910] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.209910] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.209910] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.209910] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.209910] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.209910] env[61964]: ERROR nova.compute.manager raise self.value [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.209910] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.209910] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.209910] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.210369] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.210369] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.210369] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 617.210369] env[61964]: ERROR nova.compute.manager [ 617.210369] env[61964]: Traceback (most recent call last): [ 617.210369] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.210369] env[61964]: listener.cb(fileno) [ 617.210369] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.210369] env[61964]: result = function(*args, **kwargs) [ 617.210369] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.210369] env[61964]: return func(*args, **kwargs) [ 617.210369] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.210369] env[61964]: raise e [ 617.210369] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.210369] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 617.210369] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.210369] env[61964]: created_port_ids = self._update_ports_for_instance( [ 617.210369] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.210369] env[61964]: with excutils.save_and_reraise_exception(): [ 617.210369] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.210369] env[61964]: self.force_reraise() [ 617.210369] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.210369] env[61964]: raise self.value [ 617.210369] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.210369] env[61964]: updated_port = self._update_port( [ 617.210369] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.210369] env[61964]: _ensure_no_port_binding_failure(port) [ 617.210369] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.210369] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.210972] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 617.210972] env[61964]: Removing descriptor: 18 [ 617.211349] env[61964]: ERROR nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Traceback (most recent call last): [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] yield resources [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.driver.spawn(context, instance, image_meta, [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] vm_ref = self.build_virtual_machine(instance, [ 617.211349] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] for vif in network_info: [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self._sync_wrapper(fn, *args, **kwargs) [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.wait() [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self[:] = self._gt.wait() [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self._exit_event.wait() [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.211632] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] result = hub.switch() [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self.greenlet.switch() [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] result = function(*args, **kwargs) [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return func(*args, **kwargs) [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise e [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] nwinfo = self.network_api.allocate_for_instance( [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] created_port_ids = self._update_ports_for_instance( [ 617.211911] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] with excutils.save_and_reraise_exception(): [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.force_reraise() [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise self.value [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] updated_port = self._update_port( [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] _ensure_no_port_binding_failure(port) [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise exception.PortBindingFailed(port_id=port['id']) [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 617.212226] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] [ 617.212579] env[61964]: INFO nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Terminating instance [ 617.278422] env[61964]: DEBUG nova.compute.manager [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Received event network-changed-2b507ba0-2a0c-44c7-8ebe-64cbb9c71592 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 617.278476] env[61964]: DEBUG nova.compute.manager [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Refreshing instance network info cache due to event network-changed-2b507ba0-2a0c-44c7-8ebe-64cbb9c71592. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 617.278984] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] Acquiring lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.278984] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] Acquired lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.278984] env[61964]: DEBUG nova.network.neutron [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Refreshing network info cache for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 617.368213] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.396346] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 617.397035] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 617.443665] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69ed4cb-0db9-4482-837d-db638a4b81a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.454555] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69730028-e4ec-471e-ac93-25a70cf855f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.492343] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f902149-29ef-4790-9ef5-20da9d29a413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.501348] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4188a9-0985-4f14-9136-30a9bf3bebf2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.516542] env[61964]: DEBUG nova.compute.provider_tree [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.717340] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.806045] env[61964]: DEBUG nova.network.neutron [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.871069] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Releasing lock "refresh_cache-19a6958e-4887-459b-8bf7-dd4cbbedb700" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.871311] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 617.871728] env[61964]: DEBUG nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 617.871951] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.896580] env[61964]: DEBUG nova.network.neutron [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.898759] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.907205] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 617.907205] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 617.907205] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 618.020917] env[61964]: DEBUG nova.scheduler.client.report [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 618.293729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "6ee845c3-5f79-4704-8b7b-cd3770202647" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.293995] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.400610] env[61964]: DEBUG oslo_concurrency.lockutils [req-8ad86e4f-235f-40c4-9aa8-9a37e47ebffd req-69fda6eb-0985-4d6e-b48e-b1814b523751 service nova] Releasing lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.401103] env[61964]: DEBUG nova.network.neutron [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.402345] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquired lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.402646] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.412387] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 618.412545] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 618.412672] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 618.412801] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Didn't find any instances for network info cache update. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10456}} [ 618.413575] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.413748] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.413896] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.414056] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.414205] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.414800] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.414919] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 618.415903] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.527384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.528038] env[61964]: ERROR nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Traceback (most recent call last): [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.driver.spawn(context, instance, image_meta, [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] vm_ref = self.build_virtual_machine(instance, [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.528038] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] for vif in network_info: [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self._sync_wrapper(fn, *args, **kwargs) [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.wait() [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self[:] = self._gt.wait() [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self._exit_event.wait() [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] result = hub.switch() [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.528360] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return self.greenlet.switch() [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] result = function(*args, **kwargs) [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] return func(*args, **kwargs) [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise e [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] nwinfo = self.network_api.allocate_for_instance( [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] created_port_ids = self._update_ports_for_instance( [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] with excutils.save_and_reraise_exception(): [ 618.528738] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] self.force_reraise() [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise self.value [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] updated_port = self._update_port( [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] _ensure_no_port_binding_failure(port) [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] raise exception.PortBindingFailed(port_id=port['id']) [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] nova.exception.PortBindingFailed: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. [ 618.529117] env[61964]: ERROR nova.compute.manager [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] [ 618.529640] env[61964]: DEBUG nova.compute.utils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.530728] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.715s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.532200] env[61964]: INFO nova.compute.claims [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.534832] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Build of instance 6c45bf0b-17c2-4065-94d9-52ee7b9af151 was re-scheduled: Binding failed for port 7e708664-b5e3-4342-aebf-067a4037a464, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 618.535382] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 618.535976] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.536188] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquired lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.536360] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.908142] env[61964]: INFO nova.compute.manager [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] [instance: 19a6958e-4887-459b-8bf7-dd4cbbedb700] Took 1.03 seconds to deallocate network for instance. [ 618.925122] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.948405] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.064411] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.093423] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.283427] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.317327] env[61964]: DEBUG nova.compute.manager [req-1db452c2-a762-4b31-8cab-5f5af7da0b67 req-0ab30e5f-948f-4087-a3ea-c3b0b303b725 service nova] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Received event network-vif-deleted-2b507ba0-2a0c-44c7-8ebe-64cbb9c71592 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 619.566213] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Releasing lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.568144] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 619.568144] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 619.568144] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14f00109-5ac4-44e0-a98c-7a32c2bdf086 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.579436] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b41c496-dd70-4845-baf3-378dfe676aa7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.613097] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1167bfd6-ba4f-4b19-bea6-e934bc61829d could not be found. [ 619.614343] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 619.614343] env[61964]: INFO nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 619.614343] env[61964]: DEBUG oslo.service.loopingcall [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.614343] env[61964]: DEBUG nova.compute.manager [-] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 619.614343] env[61964]: DEBUG nova.network.neutron [-] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.649414] env[61964]: DEBUG nova.network.neutron [-] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.788476] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Releasing lock "refresh_cache-6c45bf0b-17c2-4065-94d9-52ee7b9af151" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.789268] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 619.789268] env[61964]: DEBUG nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 619.789268] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.831779] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.951405] env[61964]: INFO nova.scheduler.client.report [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Deleted allocations for instance 19a6958e-4887-459b-8bf7-dd4cbbedb700 [ 620.029567] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee5c35b-dc8a-43ed-acb1-3de6145c1578 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.040645] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8a20e6-ec92-4e50-be6f-dfbb37215897 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.071787] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dbefd6-e640-404b-a0b1-fde200160240 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.080768] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bc38ec-5b9a-4538-a1b9-d0f24f10b6f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.094913] env[61964]: DEBUG nova.compute.provider_tree [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.150944] env[61964]: DEBUG nova.network.neutron [-] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.335599] env[61964]: DEBUG nova.network.neutron [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.464885] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4c0ea213-1db6-4531-8afe-c4429ec50364 tempest-AttachInterfacesUnderV243Test-1766956023 tempest-AttachInterfacesUnderV243Test-1766956023-project-member] Lock "19a6958e-4887-459b-8bf7-dd4cbbedb700" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.090s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.599357] env[61964]: DEBUG nova.scheduler.client.report [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 620.653999] env[61964]: INFO nova.compute.manager [-] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Took 1.04 seconds to deallocate network for instance. [ 620.656359] env[61964]: DEBUG nova.compute.claims [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.656525] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.841033] env[61964]: INFO nova.compute.manager [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 6c45bf0b-17c2-4065-94d9-52ee7b9af151] Took 1.05 seconds to deallocate network for instance. [ 620.971219] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 621.107105] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.107664] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 621.114407] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.358s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.511266] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.626597] env[61964]: DEBUG nova.compute.utils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.628342] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 621.628713] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.706712] env[61964]: DEBUG nova.policy [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '731ab2c060944f41b8bd293b3282a703', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912fe686053f4633b1099dce01d72365', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.885964] env[61964]: INFO nova.scheduler.client.report [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Deleted allocations for instance 6c45bf0b-17c2-4065-94d9-52ee7b9af151 [ 622.132141] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 622.185089] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591b89e3-4ee5-4426-865c-03967e2bcbe6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.196559] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8472c90e-4537-4b90-b080-3199bf222360 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.236730] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424eec33-a263-4225-af5c-ad473702104f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.245250] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c67f438-f7ef-421a-bebf-871b9d5a5f34 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.259155] env[61964]: DEBUG nova.compute.provider_tree [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.399280] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8501a03d-b05c-4e37-a72d-da929a284119 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "6c45bf0b-17c2-4065-94d9-52ee7b9af151" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.763s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.471865] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Successfully created port: 73b94651-31bb-421c-a2ce-9457cd2f5dd2 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.764967] env[61964]: DEBUG nova.scheduler.client.report [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.904316] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 623.149653] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 623.174610] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Successfully created port: a6167722-f3e2-4351-b9aa-33b0e342cf39 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.191827] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 623.192059] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.192219] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.192394] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.193474] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.193474] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 623.193474] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 623.193614] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 623.193745] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 623.193900] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 623.194086] env[61964]: DEBUG nova.virt.hardware [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 623.195042] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf91767-560d-492c-a81b-1fcbabafa92a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.204273] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad390ca-61d0-403c-be8b-fb6062b9e0aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.272334] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.157s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.272334] env[61964]: ERROR nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Traceback (most recent call last): [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.driver.spawn(context, instance, image_meta, [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.272334] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] vm_ref = self.build_virtual_machine(instance, [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] for vif in network_info: [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self._sync_wrapper(fn, *args, **kwargs) [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.wait() [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self[:] = self._gt.wait() [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self._exit_event.wait() [ 623.272567] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] result = hub.switch() [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return self.greenlet.switch() [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] result = function(*args, **kwargs) [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] return func(*args, **kwargs) [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise e [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] nwinfo = self.network_api.allocate_for_instance( [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.272856] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] created_port_ids = self._update_ports_for_instance( [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] with excutils.save_and_reraise_exception(): [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] self.force_reraise() [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise self.value [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] updated_port = self._update_port( [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] _ensure_no_port_binding_failure(port) [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.273177] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] raise exception.PortBindingFailed(port_id=port['id']) [ 623.273451] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] nova.exception.PortBindingFailed: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. [ 623.273451] env[61964]: ERROR nova.compute.manager [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] [ 623.273451] env[61964]: DEBUG nova.compute.utils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.275069] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Build of instance 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4 was re-scheduled: Binding failed for port adf548f1-d6d1-4418-97b3-68733c9aa028, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 623.275537] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 623.275929] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquiring lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.276119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Acquired lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.276290] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.277400] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.757s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.283029] env[61964]: INFO nova.compute.claims [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 623.434405] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.811126] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.889331] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.937308] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Successfully created port: b831f9ee-2a06-4061-aaf2-906c4260521f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.394567] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Releasing lock "refresh_cache-4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.394567] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 624.394567] env[61964]: DEBUG nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 624.394567] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.436515] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.545779] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.546018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.788918] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966735e4-9e0d-4a72-b8d9-5e674cd783b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.797175] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7783353-f344-4524-b9b5-966132afd0b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.829633] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d747cbe5-9329-4338-bc6f-6752c343da5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.837856] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f73c3e7-1f90-45c2-95c2-553088235357 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.851837] env[61964]: DEBUG nova.compute.provider_tree [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.944335] env[61964]: DEBUG nova.network.neutron [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.079948] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.080219] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.359176] env[61964]: DEBUG nova.scheduler.client.report [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 625.447958] env[61964]: INFO nova.compute.manager [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] [instance: 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4] Took 1.05 seconds to deallocate network for instance. [ 625.863200] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.863804] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 625.870018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.884s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.264192] env[61964]: DEBUG nova.compute.manager [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Received event network-changed-73b94651-31bb-421c-a2ce-9457cd2f5dd2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 626.264192] env[61964]: DEBUG nova.compute.manager [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Refreshing instance network info cache due to event network-changed-73b94651-31bb-421c-a2ce-9457cd2f5dd2. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 626.264330] env[61964]: DEBUG oslo_concurrency.lockutils [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] Acquiring lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.264396] env[61964]: DEBUG oslo_concurrency.lockutils [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] Acquired lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.264555] env[61964]: DEBUG nova.network.neutron [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Refreshing network info cache for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 626.295160] env[61964]: ERROR nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 626.295160] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.295160] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.295160] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.295160] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.295160] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.295160] env[61964]: ERROR nova.compute.manager raise self.value [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.295160] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.295160] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.295160] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.295608] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.295608] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.295608] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 626.295608] env[61964]: ERROR nova.compute.manager [ 626.295608] env[61964]: Traceback (most recent call last): [ 626.295608] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.295608] env[61964]: listener.cb(fileno) [ 626.295608] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.295608] env[61964]: result = function(*args, **kwargs) [ 626.295608] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.295608] env[61964]: return func(*args, **kwargs) [ 626.295608] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.295608] env[61964]: raise e [ 626.295608] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.295608] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 626.295608] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.295608] env[61964]: created_port_ids = self._update_ports_for_instance( [ 626.295608] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.295608] env[61964]: with excutils.save_and_reraise_exception(): [ 626.295608] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.295608] env[61964]: self.force_reraise() [ 626.295608] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.295608] env[61964]: raise self.value [ 626.295608] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.295608] env[61964]: updated_port = self._update_port( [ 626.295608] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.295608] env[61964]: _ensure_no_port_binding_failure(port) [ 626.295608] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.295608] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.296356] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 626.296356] env[61964]: Removing descriptor: 17 [ 626.298342] env[61964]: ERROR nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Traceback (most recent call last): [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] yield resources [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.driver.spawn(context, instance, image_meta, [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] vm_ref = self.build_virtual_machine(instance, [ 626.298342] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] for vif in network_info: [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self._sync_wrapper(fn, *args, **kwargs) [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.wait() [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self[:] = self._gt.wait() [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self._exit_event.wait() [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.298685] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] result = hub.switch() [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self.greenlet.switch() [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] result = function(*args, **kwargs) [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return func(*args, **kwargs) [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise e [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] nwinfo = self.network_api.allocate_for_instance( [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] created_port_ids = self._update_ports_for_instance( [ 626.298993] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] with excutils.save_and_reraise_exception(): [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.force_reraise() [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise self.value [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] updated_port = self._update_port( [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] _ensure_no_port_binding_failure(port) [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise exception.PortBindingFailed(port_id=port['id']) [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 626.299321] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] [ 626.299648] env[61964]: INFO nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Terminating instance [ 626.374677] env[61964]: DEBUG nova.compute.utils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.376226] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 626.376403] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 626.447939] env[61964]: DEBUG nova.policy [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5089354afb1247edb0fc665d0e2cde55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb8d46e01f2c424cb963193986b3a049', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.481810] env[61964]: INFO nova.scheduler.client.report [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Deleted allocations for instance 4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4 [ 626.804766] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.816817] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b92096-7d63-4431-9fc2-7e702aeb249e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.824988] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759e9c70-f313-4447-b264-222e687deace {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.863701] env[61964]: DEBUG nova.network.neutron [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.866082] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d25e70-fa72-46e9-9775-fbc5071e4a5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.878491] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2953246-01af-444d-bb20-814502aaa573 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.881080] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 626.897237] env[61964]: DEBUG nova.compute.provider_tree [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.899750] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Successfully created port: 130d84f8-aab1-4adc-a914-6938a110eaf9 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.997706] env[61964]: DEBUG oslo_concurrency.lockutils [None req-05a79460-2eaf-4a5c-8dc7-536562721987 tempest-ServerExternalEventsTest-241879194 tempest-ServerExternalEventsTest-241879194-project-member] Lock "4e5d1c9d-9bea-48c3-98a6-4eb29fbecae4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.798s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.133896] env[61964]: DEBUG nova.network.neutron [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.404380] env[61964]: DEBUG nova.scheduler.client.report [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 627.501053] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 627.637936] env[61964]: DEBUG oslo_concurrency.lockutils [req-de9ecf1a-7a62-4d15-b8ec-58226d296785 req-aa75eb37-8ba0-459f-9790-defdf66d2510 service nova] Releasing lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.638421] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.638576] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.896509] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 627.913119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.913119] env[61964]: ERROR nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Traceback (most recent call last): [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.driver.spawn(context, instance, image_meta, [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.913119] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] vm_ref = self.build_virtual_machine(instance, [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] for vif in network_info: [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self._sync_wrapper(fn, *args, **kwargs) [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.wait() [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self[:] = self._gt.wait() [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self._exit_event.wait() [ 627.913425] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] result = hub.switch() [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return self.greenlet.switch() [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] result = function(*args, **kwargs) [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] return func(*args, **kwargs) [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise e [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] nwinfo = self.network_api.allocate_for_instance( [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.913822] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] created_port_ids = self._update_ports_for_instance( [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] with excutils.save_and_reraise_exception(): [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] self.force_reraise() [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise self.value [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] updated_port = self._update_port( [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] _ensure_no_port_binding_failure(port) [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.914172] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] raise exception.PortBindingFailed(port_id=port['id']) [ 627.914503] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] nova.exception.PortBindingFailed: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. [ 627.914503] env[61964]: ERROR nova.compute.manager [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] [ 627.914503] env[61964]: DEBUG nova.compute.utils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.917036] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.374s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.918082] env[61964]: INFO nova.compute.claims [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.925258] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Build of instance 669fe2d9-cb37-4757-a199-18877cc6149e was re-scheduled: Binding failed for port 5d254e55-abb4-49fe-9f38-6661ebc50141, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 627.927283] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 627.927283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquiring lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.927283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Acquired lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.927283] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.938549] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 627.939192] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.939192] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 627.939192] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.939192] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 627.939360] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 627.939794] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 627.939794] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 627.939869] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 627.940040] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 627.940705] env[61964]: DEBUG nova.virt.hardware [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 627.941454] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bfa016-4a7b-41b0-976b-2b6be871eac7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.953192] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d2752f-7bf3-4bf3-bf60-94f7b43bae52 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.031842] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.204096] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.365405] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.462404] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.559322] env[61964]: DEBUG nova.compute.manager [req-403bfaa6-9a9f-4c31-a07e-097564863745 req-9bdcb3d6-08a3-44b0-90bc-3c67a1b286a0 service nova] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Received event network-vif-deleted-73b94651-31bb-421c-a2ce-9457cd2f5dd2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 628.601892] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.867688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Releasing lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.868155] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 628.868349] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 628.868661] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ccd4842-b7a9-419e-87aa-a115f88fc00b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.877882] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d67ca36-bfc0-4f5c-a330-1a7c4f8cf578 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.900685] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64818b22-3bd5-473e-8474-40bbb40ab491 could not be found. [ 628.900967] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 628.901176] env[61964]: INFO nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Took 0.03 seconds to destroy the instance on the hypervisor. [ 628.901443] env[61964]: DEBUG oslo.service.loopingcall [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.901641] env[61964]: DEBUG nova.compute.manager [-] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.901734] env[61964]: DEBUG nova.network.neutron [-] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.018297] env[61964]: ERROR nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 629.018297] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.018297] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.018297] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.018297] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.018297] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.018297] env[61964]: ERROR nova.compute.manager raise self.value [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.018297] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.018297] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.018297] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.018669] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.018669] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.018669] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 629.018669] env[61964]: ERROR nova.compute.manager [ 629.018669] env[61964]: Traceback (most recent call last): [ 629.018669] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.018669] env[61964]: listener.cb(fileno) [ 629.018669] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.018669] env[61964]: result = function(*args, **kwargs) [ 629.018669] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.018669] env[61964]: return func(*args, **kwargs) [ 629.018669] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.018669] env[61964]: raise e [ 629.018669] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.018669] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 629.018669] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.018669] env[61964]: created_port_ids = self._update_ports_for_instance( [ 629.018669] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.018669] env[61964]: with excutils.save_and_reraise_exception(): [ 629.018669] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.018669] env[61964]: self.force_reraise() [ 629.018669] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.018669] env[61964]: raise self.value [ 629.018669] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.018669] env[61964]: updated_port = self._update_port( [ 629.018669] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.018669] env[61964]: _ensure_no_port_binding_failure(port) [ 629.018669] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.018669] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.019380] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 629.019380] env[61964]: Removing descriptor: 17 [ 629.019380] env[61964]: ERROR nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Traceback (most recent call last): [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] yield resources [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.driver.spawn(context, instance, image_meta, [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.019380] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] vm_ref = self.build_virtual_machine(instance, [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] for vif in network_info: [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self._sync_wrapper(fn, *args, **kwargs) [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.wait() [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self[:] = self._gt.wait() [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self._exit_event.wait() [ 629.019656] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] result = hub.switch() [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self.greenlet.switch() [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] result = function(*args, **kwargs) [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return func(*args, **kwargs) [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise e [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] nwinfo = self.network_api.allocate_for_instance( [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.020441] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] created_port_ids = self._update_ports_for_instance( [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] with excutils.save_and_reraise_exception(): [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.force_reraise() [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise self.value [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] updated_port = self._update_port( [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] _ensure_no_port_binding_failure(port) [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.020887] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise exception.PortBindingFailed(port_id=port['id']) [ 629.021482] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 629.021482] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] [ 629.021482] env[61964]: INFO nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Terminating instance [ 629.105070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Releasing lock "refresh_cache-669fe2d9-cb37-4757-a199-18877cc6149e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.105312] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 629.105479] env[61964]: DEBUG nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 629.105639] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.136255] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.219878] env[61964]: DEBUG nova.network.neutron [-] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.435076] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701e852f-b65e-430b-84b3-6fa790c349d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.444884] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a339920-c746-41e3-87e9-02211f225398 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.481460] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528c2fe4-4b9e-46d9-856c-c04970f9103e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.490599] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3080835-7b41-412f-b1a0-575d02f72818 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.506762] env[61964]: DEBUG nova.compute.provider_tree [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.516790] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquiring lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.516790] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquired lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.517995] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.637508] env[61964]: DEBUG nova.network.neutron [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.741652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "b50e683f-79b2-419a-ac95-94908adb2b3e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.741652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.011045] env[61964]: DEBUG nova.scheduler.client.report [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 630.043375] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.134716] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.141118] env[61964]: INFO nova.compute.manager [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] [instance: 669fe2d9-cb37-4757-a199-18877cc6149e] Took 1.03 seconds to deallocate network for instance. [ 630.401939] env[61964]: DEBUG nova.network.neutron [-] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.517645] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.517645] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 630.523015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.994s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.523419] env[61964]: INFO nova.compute.claims [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.618131] env[61964]: DEBUG nova.compute.manager [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Received event network-changed-130d84f8-aab1-4adc-a914-6938a110eaf9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 630.618131] env[61964]: DEBUG nova.compute.manager [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Refreshing instance network info cache due to event network-changed-130d84f8-aab1-4adc-a914-6938a110eaf9. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 630.618393] env[61964]: DEBUG oslo_concurrency.lockutils [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] Acquiring lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.638965] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Releasing lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.639400] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 630.639654] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 630.639928] env[61964]: DEBUG oslo_concurrency.lockutils [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] Acquired lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.640139] env[61964]: DEBUG nova.network.neutron [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Refreshing network info cache for port 130d84f8-aab1-4adc-a914-6938a110eaf9 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.641180] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52654d62-0ea1-4786-8838-2b44440c2dab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.655845] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77a2ec6-1ace-4cc8-8d71-613e746d859b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.674283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "46435c68-f85a-4360-b2b5-6296afc33c3e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.674521] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.686643] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fd4fe83c-a976-4dd8-a8a0-24f8545292e4 could not be found. [ 630.687102] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 630.687186] env[61964]: INFO nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 630.687539] env[61964]: DEBUG oslo.service.loopingcall [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.687831] env[61964]: DEBUG nova.compute.manager [-] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.687968] env[61964]: DEBUG nova.network.neutron [-] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.710465] env[61964]: DEBUG nova.network.neutron [-] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.908189] env[61964]: INFO nova.compute.manager [-] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Took 2.00 seconds to deallocate network for instance. [ 630.909092] env[61964]: DEBUG nova.compute.claims [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.909288] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.023268] env[61964]: DEBUG nova.compute.utils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.024678] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 631.024856] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 631.093051] env[61964]: DEBUG nova.policy [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '537ebace42e04bb38e956404afb29a34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06abcad6803f4af9a347c9c2c7a30247', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.169272] env[61964]: DEBUG nova.network.neutron [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.188444] env[61964]: INFO nova.scheduler.client.report [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Deleted allocations for instance 669fe2d9-cb37-4757-a199-18877cc6149e [ 631.213037] env[61964]: DEBUG nova.network.neutron [-] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.404618] env[61964]: DEBUG nova.network.neutron [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.501152] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Successfully created port: 25dacfb1-86f4-4393-97e6-28c9aa035c50 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.531334] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 631.700788] env[61964]: DEBUG oslo_concurrency.lockutils [None req-692591fe-6f88-488e-b187-963802086883 tempest-DeleteServersAdminTestJSON-601877735 tempest-DeleteServersAdminTestJSON-601877735-project-member] Lock "669fe2d9-cb37-4757-a199-18877cc6149e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.234s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.721144] env[61964]: INFO nova.compute.manager [-] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Took 1.03 seconds to deallocate network for instance. [ 631.727891] env[61964]: DEBUG nova.compute.claims [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.728140] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.853888] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "85cd73df-a2f2-4f54-b984-832f9589a236" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.854295] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.908053] env[61964]: DEBUG oslo_concurrency.lockutils [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] Releasing lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.908576] env[61964]: DEBUG nova.compute.manager [req-02078858-75b9-4e0a-9fe2-c848fcfdbe49 req-e57446fd-3fbf-4d7c-8581-2c0dac7651f2 service nova] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Received event network-vif-deleted-130d84f8-aab1-4adc-a914-6938a110eaf9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 632.018403] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357a0a13-9f55-4a33-beb6-1aea2784a190 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.029036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e8360c-1b2a-40df-aef0-6ef1fcf24908 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.066053] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20a6eaa-7970-4f1b-a2cc-61565a6269de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.074261] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cb159a-0b0f-4806-b4b1-0ff409d1d5dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.088976] env[61964]: DEBUG nova.compute.provider_tree [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.208143] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 632.394604] env[61964]: DEBUG nova.compute.manager [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Received event network-changed-25dacfb1-86f4-4393-97e6-28c9aa035c50 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 632.395088] env[61964]: DEBUG nova.compute.manager [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Refreshing instance network info cache due to event network-changed-25dacfb1-86f4-4393-97e6-28c9aa035c50. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 632.395446] env[61964]: DEBUG oslo_concurrency.lockutils [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] Acquiring lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.395623] env[61964]: DEBUG oslo_concurrency.lockutils [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] Acquired lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.397019] env[61964]: DEBUG nova.network.neutron [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Refreshing network info cache for port 25dacfb1-86f4-4393-97e6-28c9aa035c50 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 632.567054] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 632.595184] env[61964]: DEBUG nova.scheduler.client.report [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.601871] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 632.602413] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.602682] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 632.602955] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.603219] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 632.603481] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 632.603795] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 632.604153] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 632.604415] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 632.604690] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 632.604958] env[61964]: DEBUG nova.virt.hardware [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 632.606149] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9793dc2c-7c9f-4b09-a9ef-65e9529fc53c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.618240] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2678076f-5295-4b1f-b0fd-342ed0659562 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.717552] env[61964]: ERROR nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 632.717552] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.717552] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.717552] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.717552] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.717552] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.717552] env[61964]: ERROR nova.compute.manager raise self.value [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.717552] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.717552] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.717552] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.718307] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.718307] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.718307] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 632.718307] env[61964]: ERROR nova.compute.manager [ 632.718307] env[61964]: Traceback (most recent call last): [ 632.718307] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.718307] env[61964]: listener.cb(fileno) [ 632.718307] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.718307] env[61964]: result = function(*args, **kwargs) [ 632.718307] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.718307] env[61964]: return func(*args, **kwargs) [ 632.718307] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.718307] env[61964]: raise e [ 632.718307] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.718307] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 632.718307] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.718307] env[61964]: created_port_ids = self._update_ports_for_instance( [ 632.718307] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.718307] env[61964]: with excutils.save_and_reraise_exception(): [ 632.718307] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.718307] env[61964]: self.force_reraise() [ 632.718307] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.718307] env[61964]: raise self.value [ 632.718307] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.718307] env[61964]: updated_port = self._update_port( [ 632.718307] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.718307] env[61964]: _ensure_no_port_binding_failure(port) [ 632.718307] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.718307] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.720569] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 632.720569] env[61964]: Removing descriptor: 18 [ 632.720569] env[61964]: ERROR nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Traceback (most recent call last): [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] yield resources [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.driver.spawn(context, instance, image_meta, [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.720569] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] vm_ref = self.build_virtual_machine(instance, [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] for vif in network_info: [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self._sync_wrapper(fn, *args, **kwargs) [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.wait() [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self[:] = self._gt.wait() [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self._exit_event.wait() [ 632.720888] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] result = hub.switch() [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self.greenlet.switch() [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] result = function(*args, **kwargs) [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return func(*args, **kwargs) [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise e [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] nwinfo = self.network_api.allocate_for_instance( [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.721225] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] created_port_ids = self._update_ports_for_instance( [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] with excutils.save_and_reraise_exception(): [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.force_reraise() [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise self.value [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] updated_port = self._update_port( [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] _ensure_no_port_binding_failure(port) [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.721560] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise exception.PortBindingFailed(port_id=port['id']) [ 632.721834] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 632.721834] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] [ 632.721834] env[61964]: INFO nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Terminating instance [ 632.750512] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.923547] env[61964]: DEBUG nova.network.neutron [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.080981] env[61964]: DEBUG nova.network.neutron [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.112869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.113418] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 633.116081] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.582s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.117564] env[61964]: INFO nova.compute.claims [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.238758] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.584511] env[61964]: DEBUG oslo_concurrency.lockutils [req-8cb87341-89a0-4b28-8002-f1ebf5a939e9 req-dd8a5eaf-324a-427c-a57a-8612175c3211 service nova] Releasing lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.584916] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquired lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.585115] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.622438] env[61964]: DEBUG nova.compute.utils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.623868] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 633.624053] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.689803] env[61964]: DEBUG nova.policy [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.114655] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.117227] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Successfully created port: 51874c34-0966-45ea-b1bd-5d0809a69485 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.127515] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 634.336197] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.432790] env[61964]: DEBUG nova.compute.manager [req-4afa142c-34ec-4342-b6b5-675a99b0f2c6 req-2daa15e2-d974-4319-8044-5c0b9256cbd7 service nova] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Received event network-vif-deleted-25dacfb1-86f4-4393-97e6-28c9aa035c50 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 634.578016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9682edb6-5f3a-43d7-bd94-1ce6b645f14c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.587388] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d6ed73-6731-44ba-bf16-b7f8516d8014 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.621306] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc422ee3-6149-4c7d-b81f-f62cc8165731 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.631640] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0baba78a-9126-478b-9b35-b177fbed5c51 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.650392] env[61964]: DEBUG nova.compute.provider_tree [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.837928] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Releasing lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.838964] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 634.839189] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.839501] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a649de2-8c74-4eb0-9daa-b368621afe04 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.850154] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7cc8c6-7c85-43e1-802a-d903e5072ec7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.879827] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa526175-3381-48e8-bd5c-73ed37931dbd could not be found. [ 634.880414] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 634.882015] env[61964]: INFO nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.882355] env[61964]: DEBUG oslo.service.loopingcall [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.882639] env[61964]: DEBUG nova.compute.manager [-] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.882758] env[61964]: DEBUG nova.network.neutron [-] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.911325] env[61964]: DEBUG nova.network.neutron [-] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.155120] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 635.156354] env[61964]: DEBUG nova.scheduler.client.report [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.192879] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 635.193498] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.193709] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 635.193910] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.194119] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 635.194574] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 635.194822] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 635.195075] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 635.195267] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 635.195477] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 635.195789] env[61964]: DEBUG nova.virt.hardware [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 635.196617] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c771df-cbd9-4cb4-9c94-1f1fb0823131 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.209231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd1ef91-9a19-4ca5-a411-5ebeffeb1b45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.375466] env[61964]: ERROR nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 635.375466] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.375466] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.375466] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.375466] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.375466] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.375466] env[61964]: ERROR nova.compute.manager raise self.value [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.375466] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.375466] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.375466] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.375925] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.375925] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.375925] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 635.375925] env[61964]: ERROR nova.compute.manager [ 635.375925] env[61964]: Traceback (most recent call last): [ 635.375925] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.375925] env[61964]: listener.cb(fileno) [ 635.375925] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.375925] env[61964]: result = function(*args, **kwargs) [ 635.375925] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.375925] env[61964]: return func(*args, **kwargs) [ 635.375925] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.375925] env[61964]: raise e [ 635.375925] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.375925] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 635.375925] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.375925] env[61964]: created_port_ids = self._update_ports_for_instance( [ 635.375925] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.375925] env[61964]: with excutils.save_and_reraise_exception(): [ 635.375925] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.375925] env[61964]: self.force_reraise() [ 635.375925] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.375925] env[61964]: raise self.value [ 635.375925] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.375925] env[61964]: updated_port = self._update_port( [ 635.375925] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.375925] env[61964]: _ensure_no_port_binding_failure(port) [ 635.375925] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.375925] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.376705] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 635.376705] env[61964]: Removing descriptor: 18 [ 635.376705] env[61964]: ERROR nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Traceback (most recent call last): [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] yield resources [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.driver.spawn(context, instance, image_meta, [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.376705] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] vm_ref = self.build_virtual_machine(instance, [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] for vif in network_info: [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self._sync_wrapper(fn, *args, **kwargs) [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.wait() [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self[:] = self._gt.wait() [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self._exit_event.wait() [ 635.377016] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] result = hub.switch() [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self.greenlet.switch() [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] result = function(*args, **kwargs) [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return func(*args, **kwargs) [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise e [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] nwinfo = self.network_api.allocate_for_instance( [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.381127] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] created_port_ids = self._update_ports_for_instance( [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] with excutils.save_and_reraise_exception(): [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.force_reraise() [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise self.value [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] updated_port = self._update_port( [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] _ensure_no_port_binding_failure(port) [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.381427] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise exception.PortBindingFailed(port_id=port['id']) [ 635.381712] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 635.381712] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] [ 635.381712] env[61964]: INFO nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Terminating instance [ 635.415031] env[61964]: DEBUG nova.network.neutron [-] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.662849] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.664346] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 635.667017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.742s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.667218] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.667378] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 635.667670] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.011s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.671122] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4867f04-690f-499b-bef6-424de8825a30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.683026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20087ece-47fe-4526-85a6-bf96ce41e8af {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.697695] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c98f0b-6c26-43a9-b948-dab1cbc79e61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.704480] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4cf767-d84b-4cb4-b222-b9e82aaa3112 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.737666] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181534MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 635.737820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.873193] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "4104dee9-34db-4396-9189-8dc677d4b677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.873442] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.880721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.880878] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.881073] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.917272] env[61964]: INFO nova.compute.manager [-] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Took 1.03 seconds to deallocate network for instance. [ 635.921360] env[61964]: DEBUG nova.compute.claims [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.922111] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.172680] env[61964]: DEBUG nova.compute.utils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.175414] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 636.175414] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.239458] env[61964]: DEBUG nova.policy [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '537ebace42e04bb38e956404afb29a34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06abcad6803f4af9a347c9c2c7a30247', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.413842] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.466676] env[61964]: DEBUG nova.compute.manager [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Received event network-changed-51874c34-0966-45ea-b1bd-5d0809a69485 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 636.466898] env[61964]: DEBUG nova.compute.manager [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Refreshing instance network info cache due to event network-changed-51874c34-0966-45ea-b1bd-5d0809a69485. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 636.467048] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] Acquiring lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.563230] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.565340] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Successfully created port: 4c812075-d67f-47ee-ae45-c11c0c174ad7 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.619432] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9317e78-a9dd-44d9-b0aa-8af543e44172 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.628246] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84bab0d-75dc-4b2f-b06c-0515af9b9d4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.663760] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fa6a5c-2e6b-41ae-939b-46be997449d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.673464] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0aa393e-ffe5-486b-a0ca-d95e424f0a95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.677892] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 636.689861] env[61964]: DEBUG nova.compute.provider_tree [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.071112] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.071112] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 637.071112] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 637.071112] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] Acquired lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.071112] env[61964]: DEBUG nova.network.neutron [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Refreshing network info cache for port 51874c34-0966-45ea-b1bd-5d0809a69485 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 637.071773] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c5eebf3-9339-4b22-ab23-90636bc34cab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.083257] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438a5e17-e145-47ea-83ad-40326bb68785 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.108764] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10a51687-3a7c-4293-9710-c043e754cd25 could not be found. [ 637.109156] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 637.111026] env[61964]: INFO nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Took 0.04 seconds to destroy the instance on the hypervisor. [ 637.111026] env[61964]: DEBUG oslo.service.loopingcall [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.111026] env[61964]: DEBUG nova.compute.manager [-] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 637.111026] env[61964]: DEBUG nova.network.neutron [-] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.136794] env[61964]: DEBUG nova.network.neutron [-] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.195269] env[61964]: DEBUG nova.scheduler.client.report [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.600284] env[61964]: DEBUG nova.network.neutron [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.638124] env[61964]: DEBUG nova.network.neutron [-] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.689415] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 637.699138] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.031s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.700418] env[61964]: ERROR nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Traceback (most recent call last): [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.driver.spawn(context, instance, image_meta, [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] vm_ref = self.build_virtual_machine(instance, [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.700418] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] for vif in network_info: [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self._sync_wrapper(fn, *args, **kwargs) [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.wait() [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self[:] = self._gt.wait() [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self._exit_event.wait() [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] result = hub.switch() [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.700833] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return self.greenlet.switch() [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] result = function(*args, **kwargs) [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] return func(*args, **kwargs) [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise e [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] nwinfo = self.network_api.allocate_for_instance( [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] created_port_ids = self._update_ports_for_instance( [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] with excutils.save_and_reraise_exception(): [ 637.701247] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] self.force_reraise() [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise self.value [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] updated_port = self._update_port( [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] _ensure_no_port_binding_failure(port) [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] raise exception.PortBindingFailed(port_id=port['id']) [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] nova.exception.PortBindingFailed: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. [ 637.701572] env[61964]: ERROR nova.compute.manager [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] [ 637.701960] env[61964]: DEBUG nova.compute.utils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.702871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.192s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.705289] env[61964]: INFO nova.compute.claims [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.709229] env[61964]: DEBUG nova.network.neutron [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.718189] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Build of instance 1167bfd6-ba4f-4b19-bea6-e934bc61829d was re-scheduled: Binding failed for port 2b507ba0-2a0c-44c7-8ebe-64cbb9c71592, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 637.719136] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 637.719136] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquiring lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.719136] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Acquired lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.719423] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.725009] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 637.725295] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.725457] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 637.725639] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.725937] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 637.726132] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 637.727055] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 637.727055] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 637.728947] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 637.728947] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 637.728947] env[61964]: DEBUG nova.virt.hardware [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 637.731369] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3252708b-9068-4d56-8f2c-ae80252d29c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.740526] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb51893d-278f-4b3b-afde-8a865ca8070e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.030742] env[61964]: ERROR nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 638.030742] env[61964]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.030742] env[61964]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.030742] env[61964]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.030742] env[61964]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.030742] env[61964]: ERROR nova.compute.manager self.force_reraise() [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.030742] env[61964]: ERROR nova.compute.manager raise self.value [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.030742] env[61964]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.030742] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.030742] env[61964]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.031422] env[61964]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.031422] env[61964]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.031422] env[61964]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 638.031422] env[61964]: ERROR nova.compute.manager [ 638.031422] env[61964]: Traceback (most recent call last): [ 638.031422] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.031422] env[61964]: listener.cb(fileno) [ 638.031422] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.031422] env[61964]: result = function(*args, **kwargs) [ 638.031422] env[61964]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.031422] env[61964]: return func(*args, **kwargs) [ 638.031422] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.031422] env[61964]: raise e [ 638.031422] env[61964]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.031422] env[61964]: nwinfo = self.network_api.allocate_for_instance( [ 638.031422] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.031422] env[61964]: created_port_ids = self._update_ports_for_instance( [ 638.031422] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.031422] env[61964]: with excutils.save_and_reraise_exception(): [ 638.031422] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.031422] env[61964]: self.force_reraise() [ 638.031422] env[61964]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.031422] env[61964]: raise self.value [ 638.031422] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.031422] env[61964]: updated_port = self._update_port( [ 638.031422] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.031422] env[61964]: _ensure_no_port_binding_failure(port) [ 638.031422] env[61964]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.031422] env[61964]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.032110] env[61964]: nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 638.032110] env[61964]: Removing descriptor: 18 [ 638.032337] env[61964]: ERROR nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Traceback (most recent call last): [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] yield resources [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.driver.spawn(context, instance, image_meta, [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] vm_ref = self.build_virtual_machine(instance, [ 638.032337] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] for vif in network_info: [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self._sync_wrapper(fn, *args, **kwargs) [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.wait() [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self[:] = self._gt.wait() [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self._exit_event.wait() [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.032648] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] result = hub.switch() [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self.greenlet.switch() [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] result = function(*args, **kwargs) [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return func(*args, **kwargs) [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise e [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] nwinfo = self.network_api.allocate_for_instance( [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] created_port_ids = self._update_ports_for_instance( [ 638.032949] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] with excutils.save_and_reraise_exception(): [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.force_reraise() [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise self.value [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] updated_port = self._update_port( [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] _ensure_no_port_binding_failure(port) [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise exception.PortBindingFailed(port_id=port['id']) [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 638.033490] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] [ 638.033849] env[61964]: INFO nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Terminating instance [ 638.140614] env[61964]: INFO nova.compute.manager [-] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Took 1.03 seconds to deallocate network for instance. [ 638.143059] env[61964]: DEBUG nova.compute.claims [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.143332] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.217343] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] Releasing lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.217604] env[61964]: DEBUG nova.compute.manager [req-e0d778a3-726f-447a-aea1-5ff986361ff3 req-001664b4-553c-4b94-96db-e1f09b4b2cfe service nova] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Received event network-vif-deleted-51874c34-0966-45ea-b1bd-5d0809a69485 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 638.241048] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.336710] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.494953] env[61964]: DEBUG nova.compute.manager [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Received event network-changed-4c812075-d67f-47ee-ae45-c11c0c174ad7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 638.494953] env[61964]: DEBUG nova.compute.manager [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Refreshing instance network info cache due to event network-changed-4c812075-d67f-47ee-ae45-c11c0c174ad7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 638.495232] env[61964]: DEBUG oslo_concurrency.lockutils [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] Acquiring lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.495319] env[61964]: DEBUG oslo_concurrency.lockutils [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] Acquired lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.495500] env[61964]: DEBUG nova.network.neutron [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Refreshing network info cache for port 4c812075-d67f-47ee-ae45-c11c0c174ad7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 638.536956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.839091] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Releasing lock "refresh_cache-1167bfd6-ba4f-4b19-bea6-e934bc61829d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.839193] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 638.839313] env[61964]: DEBUG nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 638.839473] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.859195] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.038984] env[61964]: DEBUG nova.network.neutron [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.184238] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8211bf17-dca9-4ff4-b303-f74bb4d4f246 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.192775] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861558c9-0380-4bb8-b011-50f329bd2ef8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.197557] env[61964]: DEBUG nova.network.neutron [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.224943] env[61964]: DEBUG oslo_concurrency.lockutils [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] Releasing lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.224943] env[61964]: DEBUG nova.compute.manager [req-049e40af-83f3-4e68-8672-1052a874ab88 req-6dccea63-c2e5-45c5-9efe-d7c028aa594b service nova] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Received event network-vif-deleted-4c812075-d67f-47ee-ae45-c11c0c174ad7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 639.224943] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91e2a5a-1915-433f-a460-4d62f8f199a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.227568] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquired lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.227749] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.235809] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0def66dd-e5e7-48c9-8743-c0494f0029d1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.250783] env[61964]: DEBUG nova.compute.provider_tree [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.368249] env[61964]: DEBUG nova.network.neutron [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.747856] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.753933] env[61964]: DEBUG nova.scheduler.client.report [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 639.849161] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.871448] env[61964]: INFO nova.compute.manager [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] [instance: 1167bfd6-ba4f-4b19-bea6-e934bc61829d] Took 1.03 seconds to deallocate network for instance. [ 640.258585] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.259154] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 640.262912] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.829s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.264389] env[61964]: INFO nova.compute.claims [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.351602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Releasing lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.352309] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 640.352583] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 640.353350] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbc0c508-2b26-4767-abba-c7ee770fd6f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.361998] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3d3ee9-1c1a-4309-9e2a-f73b9d13016b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.386054] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a26e916-3d1e-4831-8351-0c12fc775398 could not be found. [ 640.386054] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.386054] env[61964]: INFO nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Took 0.03 seconds to destroy the instance on the hypervisor. [ 640.386054] env[61964]: DEBUG oslo.service.loopingcall [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.386413] env[61964]: DEBUG nova.compute.manager [-] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 640.386598] env[61964]: DEBUG nova.network.neutron [-] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.408730] env[61964]: DEBUG nova.network.neutron [-] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.745800] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.747030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.772300] env[61964]: DEBUG nova.compute.utils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.774157] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 640.774282] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 640.832469] env[61964]: DEBUG nova.policy [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4afcc5d5d7bb496d9c2879bbf9d9f9ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3466856b61a44a6a9eb15e4be519fee4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.897027] env[61964]: INFO nova.scheduler.client.report [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Deleted allocations for instance 1167bfd6-ba4f-4b19-bea6-e934bc61829d [ 640.914921] env[61964]: DEBUG nova.network.neutron [-] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.113471] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Successfully created port: 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.276381] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 641.406162] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ec881064-654e-4415-93eb-e62c77956a31 tempest-ListImageFiltersTestJSON-222702463 tempest-ListImageFiltersTestJSON-222702463-project-member] Lock "1167bfd6-ba4f-4b19-bea6-e934bc61829d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.780s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.417092] env[61964]: INFO nova.compute.manager [-] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Took 1.03 seconds to deallocate network for instance. [ 641.419398] env[61964]: DEBUG nova.compute.claims [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Aborting claim: {{(pid=61964) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.419587] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.660911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81dd13a0-56f4-4da9-b522-9e146509499b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.669471] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38be158-714a-491d-9363-47f57611fcce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.701274] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2db44c-b909-4a2e-8236-8294586bcdca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.710013] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e873fa8-d2c2-4669-a5bb-0d7ac88b631b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.722308] env[61964]: DEBUG nova.compute.provider_tree [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.910235] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 642.226417] env[61964]: DEBUG nova.scheduler.client.report [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.498100] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 642.568601] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.577438] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 642.577598] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.577806] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 642.578047] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.578230] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 642.578416] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 642.578650] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 642.578834] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 642.579041] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 642.579238] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 642.579466] env[61964]: DEBUG nova.virt.hardware [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 642.580365] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce057858-f55d-494f-af39-1fcdcf12f048 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.589273] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a421f6c-c53e-4a64-80d3-6c6988fe23d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.731336] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.731913] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 642.738015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.703s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.738015] env[61964]: INFO nova.compute.claims [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.980910] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Successfully updated port: 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 643.007400] env[61964]: DEBUG nova.compute.manager [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-vif-plugged-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 643.007631] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] Acquiring lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.007830] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.007994] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.008381] env[61964]: DEBUG nova.compute.manager [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] No waiting events found dispatching network-vif-plugged-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 643.009353] env[61964]: WARNING nova.compute.manager [req-3e3440ed-e135-4f01-9773-e1b6b8c7f7eb req-ee97868a-04b8-4f8a-8c6c-49a72acccd42 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received unexpected event network-vif-plugged-88b99d42-7c9c-4409-b173-a0c8737bb0c3 for instance with vm_state building and task_state spawning. [ 643.240392] env[61964]: DEBUG nova.compute.utils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.251145] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 643.484377] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.484534] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.484692] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.752338] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.018133] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.149233] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5ce937-18b3-4838-b83a-2a48e8c855ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.157383] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b95b855-ad1e-4f9f-81aa-a45941c6ec7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.190793] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69f0789-a4f2-4663-953e-6ad7cb2164c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.196926] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595b5a92-6da1-4549-8660-450f66afd611 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.210226] env[61964]: DEBUG nova.compute.provider_tree [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.231944] env[61964]: DEBUG nova.network.neutron [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.713174] env[61964]: DEBUG nova.scheduler.client.report [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 644.734054] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.734379] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Instance network_info: |[{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 644.735491] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:36:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '073f8535-6b3a-4d21-a754-4c975554dcbf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88b99d42-7c9c-4409-b173-a0c8737bb0c3', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 644.746828] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Creating folder: Project (3466856b61a44a6a9eb15e4be519fee4). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 644.747559] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09d1c3c8-bd3d-466b-b2b5-29880c9a7755 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.759108] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Created folder: Project (3466856b61a44a6a9eb15e4be519fee4) in parent group-v230360. [ 644.759389] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Creating folder: Instances. Parent ref: group-v230374. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 644.759548] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92f25d39-bf55-46ea-a7ad-1d622cc74b28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.767035] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 644.774558] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Created folder: Instances in parent group-v230374. [ 644.774558] env[61964]: DEBUG oslo.service.loopingcall [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.774558] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 644.774558] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d0b89a86-17aa-4b47-a184-6ed23fb4d0b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.797013] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 644.797013] env[61964]: value = "task-1040557" [ 644.797013] env[61964]: _type = "Task" [ 644.797013] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.806668] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040557, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.808810] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 644.809027] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.809181] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 644.809366] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.809508] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 644.809648] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 644.809846] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 644.811011] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 644.811011] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 644.811011] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 644.811011] env[61964]: DEBUG nova.virt.hardware [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 644.811625] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af48ef76-f61c-4a26-8151-4c2f49271581 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.819973] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c7c18f-c796-4e98-9fc2-21618095a078 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.837541] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 644.843129] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Creating folder: Project (024c802c2a124235b5ad2e8152a3e1af). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 644.843454] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44dac98b-de14-469e-9c3c-b2c5915765d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.855037] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Created folder: Project (024c802c2a124235b5ad2e8152a3e1af) in parent group-v230360. [ 644.855037] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Creating folder: Instances. Parent ref: group-v230377. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 644.855037] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f43e8299-7048-4ace-ba30-ee19446a300e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.864397] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Created folder: Instances in parent group-v230377. [ 644.864690] env[61964]: DEBUG oslo.service.loopingcall [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.865092] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 644.865171] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4aef2a55-311d-4a72-8207-7ab126eeeb06 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.881422] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 644.881422] env[61964]: value = "task-1040560" [ 644.881422] env[61964]: _type = "Task" [ 644.881422] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.889093] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040560, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.047549] env[61964]: DEBUG nova.compute.manager [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 645.047549] env[61964]: DEBUG nova.compute.manager [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing instance network info cache due to event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 645.047946] env[61964]: DEBUG oslo_concurrency.lockutils [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.048200] env[61964]: DEBUG oslo_concurrency.lockutils [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.048386] env[61964]: DEBUG nova.network.neutron [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 645.219571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.220122] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 645.229180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.317s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.312185] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040557, 'name': CreateVM_Task, 'duration_secs': 0.332466} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.312344] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 645.321435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.321435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.321435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 645.321435] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d13de5e-f4f7-46ea-b00a-c52fe2736873 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.325921] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 645.325921] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ee600e-0d38-d4d5-c756-6ce53005632b" [ 645.325921] env[61964]: _type = "Task" [ 645.325921] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.333856] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ee600e-0d38-d4d5-c756-6ce53005632b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.392702] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040560, 'name': CreateVM_Task, 'duration_secs': 0.269151} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.392883] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 645.393297] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.732623] env[61964]: DEBUG nova.compute.utils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.738019] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 645.738019] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 645.787847] env[61964]: DEBUG nova.network.neutron [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updated VIF entry in instance network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 645.788316] env[61964]: DEBUG nova.network.neutron [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.819173] env[61964]: DEBUG nova.policy [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05bfff4e5dae488593812635d1c216b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f1804a0ded74317a1e2016db18c55ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.837587] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ee600e-0d38-d4d5-c756-6ce53005632b, 'name': SearchDatastore_Task, 'duration_secs': 0.009191} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.837993] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.838339] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 645.838637] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.838840] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.839091] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 645.841708] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.842080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 645.842393] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-677ea2ee-ba06-45f6-8e0e-c2556ba90369 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.844403] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0813bbd9-c7ce-4793-90aa-ec871781188c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.849283] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 645.849283] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52beab3a-42fa-9be5-59a5-6dab57fd38a6" [ 645.849283] env[61964]: _type = "Task" [ 645.849283] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.853409] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 645.853658] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 645.854695] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb3b675b-43f9-4f09-bfa9-5441f9ceb0c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.861878] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52beab3a-42fa-9be5-59a5-6dab57fd38a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.865245] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 645.865245] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529d6339-411d-d85e-ff74-146cf7c31cf4" [ 645.865245] env[61964]: _type = "Task" [ 645.865245] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.872403] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529d6339-411d-d85e-ff74-146cf7c31cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.235393] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Successfully created port: b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.237692] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 646.276065] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72da990-c743-4137-872c-ff6533c65b99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.284158] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3bab3b-bc8b-418f-9ee7-8e2a3e05d592 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.320347] env[61964]: DEBUG oslo_concurrency.lockutils [req-92590729-0ebf-4a35-b134-287619b3e081 req-574c74df-0911-4124-bb71-29323c8ca2ed service nova] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.322458] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e53e93-6bf8-4536-b012-0f4982e7d656 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.331058] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f037bbb-55ec-4f7b-8512-9c719d59b89b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.350406] env[61964]: DEBUG nova.compute.provider_tree [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.363284] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52beab3a-42fa-9be5-59a5-6dab57fd38a6, 'name': SearchDatastore_Task, 'duration_secs': 0.0083} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.364438] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.364691] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 646.364925] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.374255] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529d6339-411d-d85e-ff74-146cf7c31cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.008355} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.374990] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57dd0779-1727-4aa2-bff2-aa3611a6154c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.380020] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 646.380020] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2a715-3e71-3853-62b4-0c74cb06af7a" [ 646.380020] env[61964]: _type = "Task" [ 646.380020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.387849] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2a715-3e71-3853-62b4-0c74cb06af7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.757871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "fb538d67-47c3-4639-a642-b3358f97df31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.758344] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.860036] env[61964]: DEBUG nova.scheduler.client.report [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 646.894704] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2a715-3e71-3853-62b4-0c74cb06af7a, 'name': SearchDatastore_Task, 'duration_secs': 0.008632} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.894704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.894704] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/d8d12f21-5af8-4156-8e89-3f719199ca0a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 646.894704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.895014] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 646.895014] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b5ffdc9-9ecb-448a-9e28-76e9c7b1c249 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.897368] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1106c355-ea11-454e-9e97-b06c37f5337e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.903562] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 646.903562] env[61964]: value = "task-1040561" [ 646.903562] env[61964]: _type = "Task" [ 646.903562] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.911019] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 646.911019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 646.911019] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5cc29c5-fd5d-4ce7-b782-643b5c1df2c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.913295] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040561, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.916255] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 646.916255] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d55ab-cb38-9b3f-81a7-25e95490cf4d" [ 646.916255] env[61964]: _type = "Task" [ 646.916255] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.923230] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d55ab-cb38-9b3f-81a7-25e95490cf4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.255450] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 647.284732] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 647.285046] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.285208] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 647.285404] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.285546] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 647.285719] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 647.285894] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 647.286069] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 647.286241] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 647.286407] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 647.286589] env[61964]: DEBUG nova.virt.hardware [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 647.288764] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe7fc65-da12-4d8e-b83f-e028860e02df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.296085] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8193aac-b1c1-46c0-84ef-14ffab995fc2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.364781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.137s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.364781] env[61964]: ERROR nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Traceback (most recent call last): [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.driver.spawn(context, instance, image_meta, [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.364781] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] vm_ref = self.build_virtual_machine(instance, [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] for vif in network_info: [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self._sync_wrapper(fn, *args, **kwargs) [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.wait() [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self[:] = self._gt.wait() [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self._exit_event.wait() [ 647.365140] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] result = hub.switch() [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return self.greenlet.switch() [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] result = function(*args, **kwargs) [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] return func(*args, **kwargs) [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise e [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] nwinfo = self.network_api.allocate_for_instance( [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.365467] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] created_port_ids = self._update_ports_for_instance( [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] with excutils.save_and_reraise_exception(): [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] self.force_reraise() [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise self.value [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] updated_port = self._update_port( [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] _ensure_no_port_binding_failure(port) [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.365799] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] raise exception.PortBindingFailed(port_id=port['id']) [ 647.366216] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] nova.exception.PortBindingFailed: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. [ 647.366216] env[61964]: ERROR nova.compute.manager [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] [ 647.366216] env[61964]: DEBUG nova.compute.utils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 647.369219] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.638s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.369663] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Build of instance 64818b22-3bd5-473e-8474-40bbb40ab491 was re-scheduled: Binding failed for port 73b94651-31bb-421c-a2ce-9457cd2f5dd2, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 647.370622] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 647.370622] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.370622] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.370763] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.420655] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040561, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46185} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.422386] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/d8d12f21-5af8-4156-8e89-3f719199ca0a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 647.422855] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 647.424211] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed328dfd-4a6e-4588-a615-ec418cc29581 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.432120] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d55ab-cb38-9b3f-81a7-25e95490cf4d, 'name': SearchDatastore_Task, 'duration_secs': 0.00803} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.435079] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 647.435079] env[61964]: value = "task-1040562" [ 647.435079] env[61964]: _type = "Task" [ 647.435079] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.435079] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33f3093f-e452-4b30-a44a-68aab0809d15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.443986] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 647.443986] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c7487-a7c1-0171-f10e-6d24a2e2ea75" [ 647.443986] env[61964]: _type = "Task" [ 647.443986] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.447600] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040562, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.459636] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c7487-a7c1-0171-f10e-6d24a2e2ea75, 'name': SearchDatastore_Task, 'duration_secs': 0.009133} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.460023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.460523] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 9f216f01-dc9f-459f-88be-3a5439d3701a/9f216f01-dc9f-459f-88be-3a5439d3701a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 647.460728] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29e4fb5d-f393-4c8d-9983-b779b99c061d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.468431] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 647.468431] env[61964]: value = "task-1040563" [ 647.468431] env[61964]: _type = "Task" [ 647.468431] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.477307] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.900286] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.933331] env[61964]: DEBUG nova.compute.manager [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Received event network-vif-plugged-b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 647.933588] env[61964]: DEBUG oslo_concurrency.lockutils [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] Acquiring lock "0f1ed823-1531-4505-a68a-cb172f804655-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.933805] env[61964]: DEBUG oslo_concurrency.lockutils [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] Lock "0f1ed823-1531-4505-a68a-cb172f804655-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.933973] env[61964]: DEBUG oslo_concurrency.lockutils [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] Lock "0f1ed823-1531-4505-a68a-cb172f804655-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.934155] env[61964]: DEBUG nova.compute.manager [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] No waiting events found dispatching network-vif-plugged-b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 647.934319] env[61964]: WARNING nova.compute.manager [req-84d0c927-875e-405a-98c5-9caa9d548ff5 req-947afd33-b2ca-472c-a526-13743cdb0474 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Received unexpected event network-vif-plugged-b85285af-4413-4095-940a-792b6a5377bc for instance with vm_state building and task_state spawning. [ 647.946813] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040562, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06459} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.949521] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 647.952497] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a526833a-b726-43ee-9b4c-de5a93551429 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.978391] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/d8d12f21-5af8-4156-8e89-3f719199ca0a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 647.982284] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c97a839b-c19d-44c7-8913-da82da3eee58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.015715] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438406} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.016618] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.018703] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 9f216f01-dc9f-459f-88be-3a5439d3701a/9f216f01-dc9f-459f-88be-3a5439d3701a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 648.018925] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 648.019242] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 648.019242] env[61964]: value = "task-1040564" [ 648.019242] env[61964]: _type = "Task" [ 648.019242] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.022049] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4bce153-cb27-49b1-8ea9-6d2ee58b4bdf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.033635] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.035226] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 648.035226] env[61964]: value = "task-1040565" [ 648.035226] env[61964]: _type = "Task" [ 648.035226] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.046509] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.094535] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Successfully updated port: b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 648.439373] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ca1b6c-c47e-4fa4-a537-ad0df7560ec4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.447371] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de9f201-a451-4cb2-a800-12792a4c8f1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.477268] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed376493-36cd-41fe-a2eb-1308d4ea9cd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.484986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62785604-8684-4366-99a3-1b809b7af2c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.497826] env[61964]: DEBUG nova.compute.provider_tree [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.522730] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Releasing lock "refresh_cache-64818b22-3bd5-473e-8474-40bbb40ab491" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.523361] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 648.523361] env[61964]: DEBUG nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 648.523361] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.536288] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040564, 'name': ReconfigVM_Task, 'duration_secs': 0.276802} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.539250] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Reconfigured VM instance instance-0000001c to attach disk [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/d8d12f21-5af8-4156-8e89-3f719199ca0a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 648.540342] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57088d2a-ea43-4dd7-b3ba-f0f4a8de6a63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.542112] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.547415] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065843} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.548519] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 648.548852] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 648.548852] env[61964]: value = "task-1040566" [ 648.548852] env[61964]: _type = "Task" [ 648.548852] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.549977] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6856104f-d3a2-4c28-8f07-917b4d078db0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.572019] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] 9f216f01-dc9f-459f-88be-3a5439d3701a/9f216f01-dc9f-459f-88be-3a5439d3701a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 648.575468] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef13270b-b09d-4486-a133-e47a55919e48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.591436] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040566, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.596615] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 648.596615] env[61964]: value = "task-1040567" [ 648.596615] env[61964]: _type = "Task" [ 648.596615] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.599679] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.599809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.599951] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.606346] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040567, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.002057] env[61964]: DEBUG nova.scheduler.client.report [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 649.044661] env[61964]: DEBUG nova.network.neutron [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.062597] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040566, 'name': Rename_Task, 'duration_secs': 0.134848} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.062873] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 649.063122] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7228aff-efc8-4325-aefb-a0bffe30f049 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.069824] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 649.069824] env[61964]: value = "task-1040568" [ 649.069824] env[61964]: _type = "Task" [ 649.069824] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.078053] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.107826] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040567, 'name': ReconfigVM_Task, 'duration_secs': 0.273029} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.108499] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Reconfigured VM instance instance-0000001d to attach disk [datastore1] 9f216f01-dc9f-459f-88be-3a5439d3701a/9f216f01-dc9f-459f-88be-3a5439d3701a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 649.110981] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d0617bc-76c4-45bf-8647-400f4fb9f23e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.115880] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 649.115880] env[61964]: value = "task-1040569" [ 649.115880] env[61964]: _type = "Task" [ 649.115880] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.123398] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040569, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.156250] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.367626] env[61964]: DEBUG nova.network.neutron [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Updating instance_info_cache with network_info: [{"id": "b85285af-4413-4095-940a-792b6a5377bc", "address": "fa:16:3e:3f:23:0f", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb85285af-44", "ovs_interfaceid": "b85285af-4413-4095-940a-792b6a5377bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.506099] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.139s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.506763] env[61964]: ERROR nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Traceback (most recent call last): [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.driver.spawn(context, instance, image_meta, [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] vm_ref = self.build_virtual_machine(instance, [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.506763] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] for vif in network_info: [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self._sync_wrapper(fn, *args, **kwargs) [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.wait() [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self[:] = self._gt.wait() [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self._exit_event.wait() [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] result = hub.switch() [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.507122] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return self.greenlet.switch() [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] result = function(*args, **kwargs) [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] return func(*args, **kwargs) [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise e [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] nwinfo = self.network_api.allocate_for_instance( [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] created_port_ids = self._update_ports_for_instance( [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] with excutils.save_and_reraise_exception(): [ 649.507571] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] self.force_reraise() [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise self.value [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] updated_port = self._update_port( [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] _ensure_no_port_binding_failure(port) [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] raise exception.PortBindingFailed(port_id=port['id']) [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] nova.exception.PortBindingFailed: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. [ 649.507987] env[61964]: ERROR nova.compute.manager [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] [ 649.508370] env[61964]: DEBUG nova.compute.utils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.509422] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.759s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.510572] env[61964]: INFO nova.compute.claims [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.513701] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Build of instance fd4fe83c-a976-4dd8-a8a0-24f8545292e4 was re-scheduled: Binding failed for port 130d84f8-aab1-4adc-a914-6938a110eaf9, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 649.513968] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 649.514181] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquiring lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.514325] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Acquired lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.514481] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.547211] env[61964]: INFO nova.compute.manager [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 64818b22-3bd5-473e-8474-40bbb40ab491] Took 1.02 seconds to deallocate network for instance. [ 649.579218] env[61964]: DEBUG oslo_vmware.api [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040568, 'name': PowerOnVM_Task, 'duration_secs': 0.452142} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.579506] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 649.579714] env[61964]: INFO nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Took 7.08 seconds to spawn the instance on the hypervisor. [ 649.579892] env[61964]: DEBUG nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 649.580705] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5c8117-90d6-4e4c-94d8-5842c375600c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.625649] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040569, 'name': Rename_Task, 'duration_secs': 0.133512} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.625944] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 649.626194] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d297866-bbad-43a8-9cf8-2b8f59f284c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.632078] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 649.632078] env[61964]: value = "task-1040570" [ 649.632078] env[61964]: _type = "Task" [ 649.632078] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.639180] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.870715] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.871090] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance network_info: |[{"id": "b85285af-4413-4095-940a-792b6a5377bc", "address": "fa:16:3e:3f:23:0f", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb85285af-44", "ovs_interfaceid": "b85285af-4413-4095-940a-792b6a5377bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 649.871729] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:23:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b85285af-4413-4095-940a-792b6a5377bc', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 649.879349] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating folder: Project (3f1804a0ded74317a1e2016db18c55ca). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 649.879678] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c37711b9-ed66-436c-b1b1-413bd88927b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.891132] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created folder: Project (3f1804a0ded74317a1e2016db18c55ca) in parent group-v230360. [ 649.891324] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating folder: Instances. Parent ref: group-v230380. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 649.891554] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14ae6781-ca2a-43ea-b029-467f39164e5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.899731] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created folder: Instances in parent group-v230380. [ 649.900038] env[61964]: DEBUG oslo.service.loopingcall [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 649.900227] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 649.900687] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d04f6e5-724c-4421-80bf-0ca9e7e6c7c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.920052] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 649.920052] env[61964]: value = "task-1040573" [ 649.920052] env[61964]: _type = "Task" [ 649.920052] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.927308] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040573, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.030159] env[61964]: DEBUG nova.compute.manager [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Received event network-changed-b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 650.030159] env[61964]: DEBUG nova.compute.manager [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Refreshing instance network info cache due to event network-changed-b85285af-4413-4095-940a-792b6a5377bc. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 650.030159] env[61964]: DEBUG oslo_concurrency.lockutils [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] Acquiring lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.030159] env[61964]: DEBUG oslo_concurrency.lockutils [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] Acquired lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.030159] env[61964]: DEBUG nova.network.neutron [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Refreshing network info cache for port b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 650.044328] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.100432] env[61964]: INFO nova.compute.manager [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Took 28.62 seconds to build instance. [ 650.142138] env[61964]: DEBUG oslo_vmware.api [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040570, 'name': PowerOnVM_Task, 'duration_secs': 0.451311} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.143091] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.144266] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 650.144428] env[61964]: INFO nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Took 5.38 seconds to spawn the instance on the hypervisor. [ 650.144627] env[61964]: DEBUG nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 650.145830] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1add3c4a-38fe-40c4-bb2a-258083b9aec9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.430610] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040573, 'name': CreateVM_Task, 'duration_secs': 0.317926} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.430705] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 650.431348] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.431513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.431832] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 650.432093] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1c394c1-ef21-4796-9be8-56a530079a4c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.436237] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 650.436237] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5297923b-04b7-ca62-0777-ce40205eb5e1" [ 650.436237] env[61964]: _type = "Task" [ 650.436237] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.444495] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5297923b-04b7-ca62-0777-ce40205eb5e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.473547] env[61964]: INFO nova.compute.manager [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Rescuing [ 650.473820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.473974] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.474165] env[61964]: DEBUG nova.network.neutron [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.579414] env[61964]: INFO nova.scheduler.client.report [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Deleted allocations for instance 64818b22-3bd5-473e-8474-40bbb40ab491 [ 650.602693] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5dc964bc-7ec6-4e8b-950b-26f19e47516f tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.006s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.650856] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Releasing lock "refresh_cache-fd4fe83c-a976-4dd8-a8a0-24f8545292e4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.650856] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 650.650856] env[61964]: DEBUG nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 650.650856] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 650.664863] env[61964]: INFO nova.compute.manager [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Took 27.26 seconds to build instance. [ 650.670393] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.811991] env[61964]: DEBUG nova.network.neutron [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Updated VIF entry in instance network info cache for port b85285af-4413-4095-940a-792b6a5377bc. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 650.812331] env[61964]: DEBUG nova.network.neutron [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Updating instance_info_cache with network_info: [{"id": "b85285af-4413-4095-940a-792b6a5377bc", "address": "fa:16:3e:3f:23:0f", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb85285af-44", "ovs_interfaceid": "b85285af-4413-4095-940a-792b6a5377bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.935307] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de270d5-db34-4ece-9e2f-865443cbbe42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.950154] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af43fc16-15d4-4752-a86b-27ba10e46648 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.953624] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5297923b-04b7-ca62-0777-ce40205eb5e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.953988] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.954311] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 650.954607] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.954814] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.955059] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 650.956172] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7715f204-9066-4827-b784-50c9312c788d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.986372] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f05c6a9-349a-4d6b-bba1-42324b4f7924 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.997165] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 650.997498] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 651.000231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3c27f9-c640-432a-aa37-9a43adea5ea3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.005372] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0233a126-9b8d-48ca-8f3c-d4ca8825f893 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.018266] env[61964]: DEBUG nova.compute.provider_tree [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.020744] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 651.020744] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520a4eb9-4425-ba1d-8547-4003edbfd799" [ 651.020744] env[61964]: _type = "Task" [ 651.020744] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.029745] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520a4eb9-4425-ba1d-8547-4003edbfd799, 'name': SearchDatastore_Task, 'duration_secs': 0.00906} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.030643] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c3590f4-f008-4186-8587-c0f3d8122a5d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.036659] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 651.036659] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234c255-bc88-3460-c13c-e45771af0bf9" [ 651.036659] env[61964]: _type = "Task" [ 651.036659] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.047306] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234c255-bc88-3460-c13c-e45771af0bf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.089739] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8149ba5b-13e8-49aa-8b27-4a5b46569881 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "64818b22-3bd5-473e-8474-40bbb40ab491" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.424s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.107868] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 651.168503] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b07e9357-335e-49ed-b257-5c80a955b0f3 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.011s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.171672] env[61964]: DEBUG nova.network.neutron [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.264835] env[61964]: DEBUG nova.network.neutron [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.317465] env[61964]: DEBUG oslo_concurrency.lockutils [req-a19ebe68-7b3c-4e7c-81cb-7b19c1334abf req-9f582370-b53b-4881-99d4-205bd92b9ebf service nova] Releasing lock "refresh_cache-0f1ed823-1531-4505-a68a-cb172f804655" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.523235] env[61964]: DEBUG nova.scheduler.client.report [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 651.549554] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234c255-bc88-3460-c13c-e45771af0bf9, 'name': SearchDatastore_Task, 'duration_secs': 0.010701} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.549924] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.550839] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0f1ed823-1531-4505-a68a-cb172f804655/0f1ed823-1531-4505-a68a-cb172f804655.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 651.551151] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-346175bf-52d8-4015-9a2b-f04c5ff9973c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.558991] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 651.558991] env[61964]: value = "task-1040574" [ 651.558991] env[61964]: _type = "Task" [ 651.558991] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.566744] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040574, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.592881] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 651.635676] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.674029] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 651.676183] env[61964]: INFO nova.compute.manager [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] [instance: fd4fe83c-a976-4dd8-a8a0-24f8545292e4] Took 1.02 seconds to deallocate network for instance. [ 651.767810] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.028343] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.028926] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 652.032621] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.295s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.070941] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040574, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470293} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.071228] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0f1ed823-1531-4505-a68a-cb172f804655/0f1ed823-1531-4505-a68a-cb172f804655.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 652.071453] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.071709] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ac64820-a82b-4458-be3b-8b73c7d0d148 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.080521] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 652.080521] env[61964]: value = "task-1040575" [ 652.080521] env[61964]: _type = "Task" [ 652.080521] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.088651] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.126056] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.202899] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.330306] env[61964]: DEBUG nova.compute.manager [None req-db114f71-984c-47d2-bbbb-05f90b661516 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 652.331294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c2d735-fbfb-4b06-8790-cef479d990e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.536700] env[61964]: DEBUG nova.compute.utils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.542311] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 652.591642] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064324} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.591642] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 652.592570] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe7ca0b-f915-4bfc-81ae-6cb1bde27063 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.608264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "9f216f01-dc9f-459f-88be-3a5439d3701a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.608264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.608264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "9f216f01-dc9f-459f-88be-3a5439d3701a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.608723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.608723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.620813] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 0f1ed823-1531-4505-a68a-cb172f804655/0f1ed823-1531-4505-a68a-cb172f804655.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 652.621854] env[61964]: INFO nova.compute.manager [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Terminating instance [ 652.625252] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abeb706c-0d4c-4466-8c33-b2499363eeb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.646803] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 652.646803] env[61964]: value = "task-1040576" [ 652.646803] env[61964]: _type = "Task" [ 652.646803] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.656622] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040576, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.714014] env[61964]: INFO nova.scheduler.client.report [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Deleted allocations for instance fd4fe83c-a976-4dd8-a8a0-24f8545292e4 [ 652.843072] env[61964]: INFO nova.compute.manager [None req-db114f71-984c-47d2-bbbb-05f90b661516 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] instance snapshotting [ 652.843952] env[61964]: DEBUG nova.objects.instance [None req-db114f71-984c-47d2-bbbb-05f90b661516 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lazy-loading 'flavor' on Instance uuid 9f216f01-dc9f-459f-88be-3a5439d3701a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 653.044911] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 653.072859] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance aa526175-3381-48e8-bd5c-73ed37931dbd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073130] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 10a51687-3a7c-4293-9710-c043e754cd25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073307] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 3a26e916-3d1e-4831-8351-0c12fc775398 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073473] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance d8d12f21-5af8-4156-8e89-3f719199ca0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073571] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 9f216f01-dc9f-459f-88be-3a5439d3701a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073675] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 0f1ed823-1531-4505-a68a-cb172f804655 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.073791] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 06631c59-9969-4763-bfe8-472e8dcf8848 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 653.143663] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "refresh_cache-9f216f01-dc9f-459f-88be-3a5439d3701a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.143917] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquired lock "refresh_cache-9f216f01-dc9f-459f-88be-3a5439d3701a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.143980] env[61964]: DEBUG nova.network.neutron [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 653.157758] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040576, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.224130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da78356c-da4f-4858-bdc7-966fbc660a5c tempest-ServersAdminNegativeTestJSON-725825099 tempest-ServersAdminNegativeTestJSON-725825099-project-member] Lock "fd4fe83c-a976-4dd8-a8a0-24f8545292e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.545s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.313122] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 653.313434] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4689687a-beab-4388-b356-fba5ebf17301 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.320247] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 653.320247] env[61964]: value = "task-1040577" [ 653.320247] env[61964]: _type = "Task" [ 653.320247] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.334764] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040577, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.350270] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207f59b3-0614-4de1-b43e-87ae93dd7c97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.368557] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ad0a2d-d11e-4829-adef-4047a156c8a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.578361] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 63793ce6-d511-403e-8a4b-cad8c4157449 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.659338] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040576, 'name': ReconfigVM_Task, 'duration_secs': 0.691855} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.659687] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 0f1ed823-1531-4505-a68a-cb172f804655/0f1ed823-1531-4505-a68a-cb172f804655.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 653.660117] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9832e7af-d819-4d82-a41a-56c2009a9706 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.668129] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 653.668129] env[61964]: value = "task-1040578" [ 653.668129] env[61964]: _type = "Task" [ 653.668129] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.676696] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040578, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.685275] env[61964]: DEBUG nova.network.neutron [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.728615] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 653.785964] env[61964]: DEBUG nova.network.neutron [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.830777] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040577, 'name': PowerOffVM_Task, 'duration_secs': 0.408885} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.831076] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 653.831888] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b9b375-a989-442f-8f55-50257512a5d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.853779] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824a47dd-f9ad-4e95-b9d7-1be29507241f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.880076] env[61964]: DEBUG nova.compute.manager [None req-db114f71-984c-47d2-bbbb-05f90b661516 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance disappeared during snapshot {{(pid=61964) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 653.891909] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 653.892252] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d156d44b-44ea-4a15-abfe-14c63edc3659 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.901387] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 653.901387] env[61964]: value = "task-1040579" [ 653.901387] env[61964]: _type = "Task" [ 653.901387] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.913507] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 653.913740] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 653.913987] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.914166] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.914345] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 653.914593] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b7b47d5-83d2-46a6-ace5-4629eaaee3ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.922993] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 653.922993] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 653.923387] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17d0d577-0c2a-44d1-8ac0-78e19fdcadb7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.929252] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 653.929252] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521d338d-a088-7719-a31c-a9010eae3955" [ 653.929252] env[61964]: _type = "Task" [ 653.929252] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.937231] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521d338d-a088-7719-a31c-a9010eae3955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.034414] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.034729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.041758] env[61964]: DEBUG nova.compute.manager [None req-db114f71-984c-47d2-bbbb-05f90b661516 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Found 0 images (rotation: 2) {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4985}} [ 654.060289] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 654.082610] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a995ff8c-75b9-4089-ad4c-9e6baa33fe8c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.091781] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 654.092163] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.092650] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 654.092650] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.092816] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 654.092859] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 654.093045] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 654.093215] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 654.093383] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 654.093568] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 654.093816] env[61964]: DEBUG nova.virt.hardware [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 654.094629] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce41518-6270-4cae-b4ab-6c483eba9bd8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.104698] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1d00d0-6773-4143-9669-e42e07c7faed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.120397] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 654.126248] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Creating folder: Project (7aa63bbcbf204aa39e5ebeef027ef85e). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 654.126637] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59ff89e5-23ae-41f5-a30e-6aa3eec05edc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.136544] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Created folder: Project (7aa63bbcbf204aa39e5ebeef027ef85e) in parent group-v230360. [ 654.136744] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Creating folder: Instances. Parent ref: group-v230383. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 654.136981] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b817947c-bf8f-4c74-9a40-9b3a77a7d024 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.147130] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Created folder: Instances in parent group-v230383. [ 654.147415] env[61964]: DEBUG oslo.service.loopingcall [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 654.147582] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 654.147866] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c925018-b2e6-4041-9c44-55993d5e5a18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.164188] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 654.164188] env[61964]: value = "task-1040582" [ 654.164188] env[61964]: _type = "Task" [ 654.164188] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.171983] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040582, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.179640] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040578, 'name': Rename_Task, 'duration_secs': 0.265557} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.179915] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 654.180217] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d06ac975-f612-433b-a5a7-74ade2629ad8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.186731] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 654.186731] env[61964]: value = "task-1040583" [ 654.186731] env[61964]: _type = "Task" [ 654.186731] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.196098] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.256745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.288750] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Releasing lock "refresh_cache-9f216f01-dc9f-459f-88be-3a5439d3701a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.289216] env[61964]: DEBUG nova.compute.manager [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 654.289414] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 654.290388] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b000578e-0400-445c-85b8-d93608c69b9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.299160] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 654.299160] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec6b79b8-45c1-47e3-b926-41cbf76dfd6b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.306032] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 654.306032] env[61964]: value = "task-1040584" [ 654.306032] env[61964]: _type = "Task" [ 654.306032] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.315468] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.441942] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521d338d-a088-7719-a31c-a9010eae3955, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.441942] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74240398-f85c-44ef-a847-b04aa7ca1174 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.447168] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 654.447168] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5263d455-fb87-286f-3bc1-1c9307f27ce9" [ 654.447168] env[61964]: _type = "Task" [ 654.447168] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.457483] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5263d455-fb87-286f-3bc1-1c9307f27ce9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.586025] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 65566c20-6d69-471c-b098-3c30c01d9955 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.674604] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040582, 'name': CreateVM_Task, 'duration_secs': 0.354455} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.674795] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 654.675279] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.675443] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.675768] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 654.676030] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6304004-2a36-458d-9da9-a5b6736dbdfe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.680793] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 654.680793] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5286fd89-e995-258e-ef68-86414f4efa50" [ 654.680793] env[61964]: _type = "Task" [ 654.680793] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.690192] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5286fd89-e995-258e-ef68-86414f4efa50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.698081] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040583, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.816924] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040584, 'name': PowerOffVM_Task, 'duration_secs': 0.120369} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.817061] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 654.817121] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 654.817367] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96c7a4cb-6390-44a7-92fc-ce83a61ca4c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.840080] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 654.840080] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 654.840280] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Deleting the datastore file [datastore1] 9f216f01-dc9f-459f-88be-3a5439d3701a {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 654.840654] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13c7fef1-a956-4e28-8076-5414d38a2f21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.847020] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for the task: (returnval){ [ 654.847020] env[61964]: value = "task-1040586" [ 654.847020] env[61964]: _type = "Task" [ 654.847020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.854667] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.957350] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5263d455-fb87-286f-3bc1-1c9307f27ce9, 'name': SearchDatastore_Task, 'duration_secs': 0.011197} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.958274] env[61964]: DEBUG oslo_concurrency.lockutils [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.958274] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. {{(pid=61964) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 654.958274] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa2de7de-de3f-48c8-a611-82c6099ad673 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.964735] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 654.964735] env[61964]: value = "task-1040587" [ 654.964735] env[61964]: _type = "Task" [ 654.964735] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.972593] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.089336] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.191637] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5286fd89-e995-258e-ef68-86414f4efa50, 'name': SearchDatastore_Task, 'duration_secs': 0.01185} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.194860] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.195382] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 655.195382] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.195556] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.195641] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 655.195895] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d369eb88-1e70-4dc0-b037-6d0fba91d9b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.202965] env[61964]: DEBUG oslo_vmware.api [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040583, 'name': PowerOnVM_Task, 'duration_secs': 0.586631} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.202965] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 655.203119] env[61964]: INFO nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Took 7.95 seconds to spawn the instance on the hypervisor. [ 655.203230] env[61964]: DEBUG nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 655.204817] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b77eb3-4d4d-4b1b-87ba-8794e9ca04e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.207141] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 655.207301] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 655.207950] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b5a0750-34d8-4438-80f1-e511be3cc94d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.217028] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 655.217028] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52753837-cec3-7afb-9229-24e0bf0b0072" [ 655.217028] env[61964]: _type = "Task" [ 655.217028] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.224150] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52753837-cec3-7afb-9229-24e0bf0b0072, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.356919] env[61964]: DEBUG oslo_vmware.api [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Task: {'id': task-1040586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193424} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.357189] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.357366] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 655.357550] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.357757] env[61964]: INFO nova.compute.manager [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Took 1.07 seconds to destroy the instance on the hypervisor. [ 655.357951] env[61964]: DEBUG oslo.service.loopingcall [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.358995] env[61964]: DEBUG nova.compute.manager [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.358995] env[61964]: DEBUG nova.network.neutron [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.376408] env[61964]: DEBUG nova.network.neutron [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.478939] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040587, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.593118] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance d99859dc-2206-42d9-ae6d-8294fbd6942c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.728163] env[61964]: INFO nova.compute.manager [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Took 27.72 seconds to build instance. [ 655.733069] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52753837-cec3-7afb-9229-24e0bf0b0072, 'name': SearchDatastore_Task, 'duration_secs': 0.012379} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.734050] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0aab090-9d98-47d5-af3c-db536f31e7e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.739413] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 655.739413] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521c9631-1ce5-3980-b9b1-2f8c7fb325b2" [ 655.739413] env[61964]: _type = "Task" [ 655.739413] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.747381] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521c9631-1ce5-3980-b9b1-2f8c7fb325b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.879404] env[61964]: DEBUG nova.network.neutron [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.975945] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.778718} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.976272] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. [ 655.977086] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c240182d-3495-42e0-bd54-6e4c1a8dd5ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.004308] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 656.004631] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-504d1850-6e0e-4d35-a33b-4a50458560fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.022354] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 656.022354] env[61964]: value = "task-1040588" [ 656.022354] env[61964]: _type = "Task" [ 656.022354] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.030365] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.096591] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6ced6800-db29-4766-8d83-b63b50d5fcc5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.236689] env[61964]: DEBUG oslo_concurrency.lockutils [None req-dd6beb0d-860a-4bbb-a904-87c31407e437 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.331s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.251112] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521c9631-1ce5-3980-b9b1-2f8c7fb325b2, 'name': SearchDatastore_Task, 'duration_secs': 0.021991} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.251963] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.252256] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 06631c59-9969-4763-bfe8-472e8dcf8848/06631c59-9969-4763-bfe8-472e8dcf8848.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 656.252548] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7de6523-06f6-4cea-9cee-65139d6939e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.260784] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 656.260784] env[61964]: value = "task-1040589" [ 656.260784] env[61964]: _type = "Task" [ 656.260784] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.270100] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040589, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.382249] env[61964]: INFO nova.compute.manager [-] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Took 1.02 seconds to deallocate network for instance. [ 656.531947] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040588, 'name': ReconfigVM_Task, 'duration_secs': 0.329452} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.532264] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Reconfigured VM instance instance-0000001c to attach disk [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.533242] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9c8dd2-d85c-4b60-b0da-e4595704313f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.562576] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7762be4b-7b91-44e1-8391-d328c998e0a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.580128] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 656.580128] env[61964]: value = "task-1040590" [ 656.580128] env[61964]: _type = "Task" [ 656.580128] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.589157] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.600091] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance aef4c3a7-641a-4356-9187-ae4c082ccde9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.741693] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 656.771486] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040589, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.893641] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.075755] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "0f1ed823-1531-4505-a68a-cb172f804655" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.076038] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.076226] env[61964]: DEBUG nova.compute.manager [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 657.077182] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61975d05-d260-4d1c-bb54-67f1f508e15e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.087999] env[61964]: DEBUG nova.compute.manager [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 657.088370] env[61964]: DEBUG nova.objects.instance [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'flavor' on Instance uuid 0f1ed823-1531-4505-a68a-cb172f804655 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 657.093516] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040590, 'name': ReconfigVM_Task, 'duration_secs': 0.155833} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.093933] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 657.094230] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5660649b-fd57-4b04-b931-e13e8dd9fe2c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.100948] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 657.100948] env[61964]: value = "task-1040591" [ 657.100948] env[61964]: _type = "Task" [ 657.100948] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.105860] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 16276be9-b305-4d1d-afde-bc98be42687a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.112246] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040591, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.273462] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040589, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.274799] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.609132] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a2a25787-1426-4c09-892d-395e1becccc9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.614046] env[61964]: DEBUG oslo_vmware.api [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040591, 'name': PowerOnVM_Task, 'duration_secs': 0.427796} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.614532] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 657.619029] env[61964]: DEBUG nova.compute.manager [None req-47c1cd63-733d-4bcb-b64c-829e034a1508 tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 657.619029] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d14b3b5-b2a9-4eb4-ac7d-b8ab58adf6a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.773535] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040589, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.047734} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.773762] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 06631c59-9969-4763-bfe8-472e8dcf8848/06631c59-9969-4763-bfe8-472e8dcf8848.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 657.773972] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.774232] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dd63853-89ac-4ed0-993e-eb1970da7a82 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.781058] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 657.781058] env[61964]: value = "task-1040592" [ 657.781058] env[61964]: _type = "Task" [ 657.781058] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.789169] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.101426] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 658.101426] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71708774-3f8f-47d0-896f-88501a5810fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.106931] env[61964]: DEBUG oslo_vmware.api [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 658.106931] env[61964]: value = "task-1040593" [ 658.106931] env[61964]: _type = "Task" [ 658.106931] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.117178] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance af413b1d-9e7e-43d0-a39a-fb9dda9cf281 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.118051] env[61964]: DEBUG oslo_vmware.api [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040593, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.291362] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064329} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.291681] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.292623] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37c0959-a2d0-4711-b82d-04b285c346ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.314749] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 06631c59-9969-4763-bfe8-472e8dcf8848/06631c59-9969-4763-bfe8-472e8dcf8848.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.315064] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9b78955-ac90-47b5-8176-ef614c7dbcd8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.333931] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 658.333931] env[61964]: value = "task-1040594" [ 658.333931] env[61964]: _type = "Task" [ 658.333931] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.342397] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.616417] env[61964]: DEBUG oslo_vmware.api [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040593, 'name': PowerOffVM_Task, 'duration_secs': 0.242801} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.616718] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 658.616915] env[61964]: DEBUG nova.compute.manager [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 658.617972] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d321a67-0b89-4fad-b0bd-e574757514b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.621330] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 5c340420-969b-4ccb-9f9a-7a833d8f0c43 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.847090] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.124455] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance df62ed65-0a89-4f04-9b5a-f5f4214e6bd5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.128670] env[61964]: DEBUG nova.compute.manager [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 659.128853] env[61964]: DEBUG nova.compute.manager [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing instance network info cache due to event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 659.129096] env[61964]: DEBUG oslo_concurrency.lockutils [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.129277] env[61964]: DEBUG oslo_concurrency.lockutils [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.129372] env[61964]: DEBUG nova.network.neutron [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 659.136018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e713793a-1bf1-496a-9da1-d50d6e9dad40 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.060s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.345716] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040594, 'name': ReconfigVM_Task, 'duration_secs': 0.927327} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.346015] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 06631c59-9969-4763-bfe8-472e8dcf8848/06631c59-9969-4763-bfe8-472e8dcf8848.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.346652] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a81daa4-6258-4e20-93f4-92c1b86e9732 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.353660] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 659.353660] env[61964]: value = "task-1040595" [ 659.353660] env[61964]: _type = "Task" [ 659.353660] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.364311] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040595, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.634579] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance df0a9a69-bd44-4da4-ba3a-9ba241c010a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.863057] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040595, 'name': Rename_Task, 'duration_secs': 0.171038} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.863338] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 659.863584] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d8db5e5-1512-4b56-8280-42feb4448400 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.870152] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 659.870152] env[61964]: value = "task-1040596" [ 659.870152] env[61964]: _type = "Task" [ 659.870152] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.877646] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.092864] env[61964]: DEBUG nova.network.neutron [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updated VIF entry in instance network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 660.092864] env[61964]: DEBUG nova.network.neutron [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.141917] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6ee845c3-5f79-4704-8b7b-cd3770202647 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.237728] env[61964]: DEBUG nova.compute.manager [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 660.237805] env[61964]: DEBUG nova.compute.manager [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing instance network info cache due to event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 660.237991] env[61964]: DEBUG oslo_concurrency.lockutils [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.380695] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040596, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.594739] env[61964]: DEBUG oslo_concurrency.lockutils [req-5c784eaa-3859-4f84-b2d7-b4c2091acd9c req-cef532df-767b-46de-8a6f-d9d8f8c76e68 service nova] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.595246] env[61964]: DEBUG oslo_concurrency.lockutils [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.595437] env[61964]: DEBUG nova.network.neutron [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 660.644430] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 0ea19f52-b700-405f-8c3c-fbaa3e17b36a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.884318] env[61964]: DEBUG oslo_vmware.api [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040596, 'name': PowerOnVM_Task, 'duration_secs': 0.752709} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.884583] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 660.887036] env[61964]: INFO nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Took 6.82 seconds to spawn the instance on the hypervisor. [ 660.887036] env[61964]: DEBUG nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 660.887036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6195fd-776d-4d4a-8b24-5313d8843503 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.973385] env[61964]: DEBUG nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 660.974339] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d4bceb-e44d-4499-900b-2a775f005259 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.147362] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance e5e7b687-5691-4b1e-b113-2b8e83b23662 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.374614] env[61964]: DEBUG nova.network.neutron [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updated VIF entry in instance network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 661.374944] env[61964]: DEBUG nova.network.neutron [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.404573] env[61964]: INFO nova.compute.manager [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Took 28.69 seconds to build instance. [ 661.490251] env[61964]: INFO nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] instance snapshotting [ 661.490251] env[61964]: WARNING nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 661.491601] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081d4396-6220-45ef-b2e1-0ab7b2552b22 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.510968] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a0f5ed-8026-4981-bc44-950e6589a847 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.653561] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance b50e683f-79b2-419a-ac95-94908adb2b3e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.877601] env[61964]: DEBUG oslo_concurrency.lockutils [req-68d23576-334f-46d5-be2f-004c4f92cb12 req-bf35107a-9e80-42da-89d3-043afb65aea0 service nova] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.885153] env[61964]: DEBUG nova.compute.manager [None req-ee2d3c6d-bdd1-4971-8732-3c9a8567d8e2 tempest-ServerDiagnosticsV248Test-1587390753 tempest-ServerDiagnosticsV248Test-1587390753-project-admin] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 661.886642] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b701ff8-a9c8-4e88-86e0-705a16968bfb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.893398] env[61964]: INFO nova.compute.manager [None req-ee2d3c6d-bdd1-4971-8732-3c9a8567d8e2 tempest-ServerDiagnosticsV248Test-1587390753 tempest-ServerDiagnosticsV248Test-1587390753-project-admin] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Retrieving diagnostics [ 661.894384] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f455616-a9b3-4390-a6a5-2278df728f3b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.927019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-03df3b44-9438-4f1d-bdc7-89f2022f8b99 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.487s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.021785] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 662.022772] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c8872535-e274-4724-b5e0-ed3153695ae2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.030201] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 662.030201] env[61964]: value = "task-1040597" [ 662.030201] env[61964]: _type = "Task" [ 662.030201] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.039471] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040597, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.157183] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 46435c68-f85a-4360-b2b5-6296afc33c3e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.171157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.171796] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.428885] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 662.541052] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040597, 'name': CreateSnapshot_Task, 'duration_secs': 0.38762} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.541338] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 662.542151] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fe7148-5977-413e-8220-809edfd46af5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.598224] env[61964]: DEBUG nova.compute.manager [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 662.598224] env[61964]: DEBUG nova.compute.manager [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing instance network info cache due to event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 662.598224] env[61964]: DEBUG oslo_concurrency.lockutils [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.598224] env[61964]: DEBUG oslo_concurrency.lockutils [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.598224] env[61964]: DEBUG nova.network.neutron [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.662013] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 85cd73df-a2f2-4f54-b984-832f9589a236 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.954309] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.064098] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 663.064897] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3bfbe3ec-40f4-43ef-b890-ba2dc725ce2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.073469] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 663.073469] env[61964]: value = "task-1040598" [ 663.073469] env[61964]: _type = "Task" [ 663.073469] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.081526] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040598, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.165515] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 4104dee9-34db-4396-9189-8dc677d4b677 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.376048] env[61964]: DEBUG nova.network.neutron [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updated VIF entry in instance network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 663.376215] env[61964]: DEBUG nova.network.neutron [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.584299] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040598, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.668864] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 58f8d368-fa0d-4c8d-9009-e6e1c5940032 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.879349] env[61964]: DEBUG oslo_concurrency.lockutils [req-e9ddf147-f598-47d9-97b8-c38fc549df06 req-84a9819b-ef6d-4492-bef0-087416e535b3 service nova] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.084336] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040598, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.171583] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance fb538d67-47c3-4639-a642-b3358f97df31 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.171879] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 664.171923] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 664.528932] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8f967e-6832-4308-b140-b0c01ab1ce95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.536190] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c79c62c-7532-4c7b-8095-c9ff5fb2074f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.565727] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3349264-89c2-46f0-a8ef-7c4d9170372d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.572751] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f9408d-15b5-4b1e-8cd4-b3727f9f3db8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.590014] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.594183] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040598, 'name': CloneVM_Task, 'duration_secs': 1.180917} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.594656] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Created linked-clone VM from snapshot [ 664.595375] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2372e6-f54e-4f9b-bcd7-0660d5803caa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.603208] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Uploading image 21f5be2c-223a-4f43-a5bb-446e2118758b {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 664.625814] env[61964]: DEBUG nova.compute.manager [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 664.625985] env[61964]: DEBUG nova.compute.manager [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing instance network info cache due to event network-changed-88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 664.626259] env[61964]: DEBUG oslo_concurrency.lockutils [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] Acquiring lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.626428] env[61964]: DEBUG oslo_concurrency.lockutils [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] Acquired lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.626626] env[61964]: DEBUG nova.network.neutron [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Refreshing network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 664.629842] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 664.629842] env[61964]: value = "vm-230387" [ 664.629842] env[61964]: _type = "VirtualMachine" [ 664.629842] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 664.630387] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8ad24344-48ae-4fa1-aee5-81951170387f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.639050] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease: (returnval){ [ 664.639050] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526aa822-de09-526c-2590-2ffe6310536f" [ 664.639050] env[61964]: _type = "HttpNfcLease" [ 664.639050] env[61964]: } obtained for exporting VM: (result){ [ 664.639050] env[61964]: value = "vm-230387" [ 664.639050] env[61964]: _type = "VirtualMachine" [ 664.639050] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 664.639425] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the lease: (returnval){ [ 664.639425] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526aa822-de09-526c-2590-2ffe6310536f" [ 664.639425] env[61964]: _type = "HttpNfcLease" [ 664.639425] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 664.646250] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 664.646250] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526aa822-de09-526c-2590-2ffe6310536f" [ 664.646250] env[61964]: _type = "HttpNfcLease" [ 664.646250] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 664.762883] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.763161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.763439] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.763636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.763863] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.766008] env[61964]: INFO nova.compute.manager [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Terminating instance [ 665.095732] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 665.147721] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 665.147721] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526aa822-de09-526c-2590-2ffe6310536f" [ 665.147721] env[61964]: _type = "HttpNfcLease" [ 665.147721] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 665.148319] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 665.148319] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526aa822-de09-526c-2590-2ffe6310536f" [ 665.148319] env[61964]: _type = "HttpNfcLease" [ 665.148319] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 665.149064] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a07d35f-1438-4dea-90d6-3130797da6a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.158019] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 665.158019] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 665.249213] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6ea175f1-a8d7-4c6f-aae2-118a5c09d77a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.270606] env[61964]: DEBUG nova.compute.manager [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 665.270818] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 665.274524] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c200f8-8b98-47c9-a447-1b16649f0975 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.288115] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 665.288377] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30e44f7f-839e-43d5-8ebc-e6bf4745b738 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.293919] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 665.293919] env[61964]: value = "task-1040600" [ 665.293919] env[61964]: _type = "Task" [ 665.293919] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.303742] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.426242] env[61964]: DEBUG nova.network.neutron [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updated VIF entry in instance network info cache for port 88b99d42-7c9c-4409-b173-a0c8737bb0c3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 665.426613] env[61964]: DEBUG nova.network.neutron [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [{"id": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "address": "fa:16:3e:e4:36:43", "network": {"id": "f1248b5f-b502-4ab4-b103-295caf4fd4b1", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-898646961-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3466856b61a44a6a9eb15e4be519fee4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "073f8535-6b3a-4d21-a754-4c975554dcbf", "external-id": "nsx-vlan-transportzone-111", "segmentation_id": 111, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b99d42-7c", "ovs_interfaceid": "88b99d42-7c9c-4409-b173-a0c8737bb0c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.601831] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 665.602101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.570s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.602390] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.681s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.807176] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040600, 'name': PowerOffVM_Task, 'duration_secs': 0.252104} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.807176] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 665.807176] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 665.807176] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-798b2bad-df29-4888-bf9d-f665d0fc9075 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.864833] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 665.865258] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 665.865557] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Deleting the datastore file [datastore1] d8d12f21-5af8-4156-8e89-3f719199ca0a {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 665.865871] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcff9c62-841d-4674-828a-f1537ae87480 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.872805] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for the task: (returnval){ [ 665.872805] env[61964]: value = "task-1040602" [ 665.872805] env[61964]: _type = "Task" [ 665.872805] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.882385] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.931438] env[61964]: DEBUG oslo_concurrency.lockutils [req-2dc9e266-e7ca-4b4e-848e-a9505945ea4c req-143e7260-9cef-4454-9aa9-00a0d635057a service nova] Releasing lock "refresh_cache-d8d12f21-5af8-4156-8e89-3f719199ca0a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.387568] env[61964]: DEBUG oslo_vmware.api [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Task: {'id': task-1040602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.425385} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.387986] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 666.388263] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 666.388526] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 666.388783] env[61964]: INFO nova.compute.manager [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 666.389147] env[61964]: DEBUG oslo.service.loopingcall [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.389421] env[61964]: DEBUG nova.compute.manager [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 666.389546] env[61964]: DEBUG nova.network.neutron [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.616974] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692dfec9-84f9-455a-bf8b-782f42bd95e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.625147] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1314d678-f330-48d5-8797-8ccf258eff3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.659699] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c107caee-07eb-4a3f-aeab-9abbd03f4f23 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.667258] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a95fa0-b416-4f80-8158-d28fe9d875c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.682487] env[61964]: DEBUG nova.compute.provider_tree [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.764457] env[61964]: DEBUG nova.compute.manager [req-090b6799-0904-41b1-9c0c-1c5c8c116a09 req-fba9ca01-2963-4253-9e6b-1cd313a5808c service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Received event network-vif-deleted-88b99d42-7c9c-4409-b173-a0c8737bb0c3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 666.764775] env[61964]: INFO nova.compute.manager [req-090b6799-0904-41b1-9c0c-1c5c8c116a09 req-fba9ca01-2963-4253-9e6b-1cd313a5808c service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Neutron deleted interface 88b99d42-7c9c-4409-b173-a0c8737bb0c3; detaching it from the instance and deleting it from the info cache [ 666.765510] env[61964]: DEBUG nova.network.neutron [req-090b6799-0904-41b1-9c0c-1c5c8c116a09 req-fba9ca01-2963-4253-9e6b-1cd313a5808c service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.187466] env[61964]: DEBUG nova.scheduler.client.report [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 667.192235] env[61964]: DEBUG nova.network.neutron [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.268245] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e88ff43f-34a7-4af3-9af4-49dc74a5518a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.278739] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b22043b-2a06-4918-9ed1-a8b9600cf167 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.302116] env[61964]: DEBUG nova.compute.manager [req-090b6799-0904-41b1-9c0c-1c5c8c116a09 req-fba9ca01-2963-4253-9e6b-1cd313a5808c service nova] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Detach interface failed, port_id=88b99d42-7c9c-4409-b173-a0c8737bb0c3, reason: Instance d8d12f21-5af8-4156-8e89-3f719199ca0a could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 667.694880] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.695617] env[61964]: ERROR nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Traceback (most recent call last): [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.driver.spawn(context, instance, image_meta, [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] vm_ref = self.build_virtual_machine(instance, [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.695617] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] for vif in network_info: [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self._sync_wrapper(fn, *args, **kwargs) [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.wait() [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self[:] = self._gt.wait() [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self._exit_event.wait() [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] result = hub.switch() [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.695963] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return self.greenlet.switch() [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] result = function(*args, **kwargs) [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] return func(*args, **kwargs) [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise e [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] nwinfo = self.network_api.allocate_for_instance( [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] created_port_ids = self._update_ports_for_instance( [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] with excutils.save_and_reraise_exception(): [ 667.696421] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] self.force_reraise() [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise self.value [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] updated_port = self._update_port( [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] _ensure_no_port_binding_failure(port) [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] raise exception.PortBindingFailed(port_id=port['id']) [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] nova.exception.PortBindingFailed: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. [ 667.696783] env[61964]: ERROR nova.compute.manager [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] [ 667.697064] env[61964]: DEBUG nova.compute.utils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.697726] env[61964]: INFO nova.compute.manager [-] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Took 1.31 seconds to deallocate network for instance. [ 667.698062] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.555s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.703180] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Build of instance aa526175-3381-48e8-bd5c-73ed37931dbd was re-scheduled: Binding failed for port 25dacfb1-86f4-4393-97e6-28c9aa035c50, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 667.704026] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 667.704026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.704175] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquired lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.704208] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.210585] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.232811] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.359480] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.605162] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265078f5-2065-4dda-b17a-035a01376970 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.613056] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3c5515-e190-45c8-95f7-5d72d596ead0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.646916] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683c9d01-1157-43d0-bfc2-e974ba9c46f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.655279] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f144150-aa5f-403b-a363-ad939c8f18df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.668902] env[61964]: DEBUG nova.compute.provider_tree [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.862584] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Releasing lock "refresh_cache-aa526175-3381-48e8-bd5c-73ed37931dbd" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.862978] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 668.863221] env[61964]: DEBUG nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.863407] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.884307] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.173300] env[61964]: DEBUG nova.scheduler.client.report [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 669.388952] env[61964]: DEBUG nova.network.neutron [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.677231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.677854] env[61964]: ERROR nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Traceback (most recent call last): [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.driver.spawn(context, instance, image_meta, [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] vm_ref = self.build_virtual_machine(instance, [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.677854] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] for vif in network_info: [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self._sync_wrapper(fn, *args, **kwargs) [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.wait() [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self[:] = self._gt.wait() [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self._exit_event.wait() [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] result = hub.switch() [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.678261] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return self.greenlet.switch() [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] result = function(*args, **kwargs) [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] return func(*args, **kwargs) [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise e [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] nwinfo = self.network_api.allocate_for_instance( [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] created_port_ids = self._update_ports_for_instance( [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] with excutils.save_and_reraise_exception(): [ 669.678626] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] self.force_reraise() [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise self.value [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] updated_port = self._update_port( [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] _ensure_no_port_binding_failure(port) [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] raise exception.PortBindingFailed(port_id=port['id']) [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] nova.exception.PortBindingFailed: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. [ 669.679017] env[61964]: ERROR nova.compute.manager [instance: 10a51687-3a7c-4293-9710-c043e754cd25] [ 669.679331] env[61964]: DEBUG nova.compute.utils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.679858] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.260s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.683720] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Build of instance 10a51687-3a7c-4293-9710-c043e754cd25 was re-scheduled: Binding failed for port 51874c34-0966-45ea-b1bd-5d0809a69485, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 669.684231] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 669.684459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.684611] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.684775] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.891362] env[61964]: INFO nova.compute.manager [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: aa526175-3381-48e8-bd5c-73ed37931dbd] Took 1.03 seconds to deallocate network for instance. [ 670.207661] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.328042] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.608745] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2a5340-e980-4c97-b61e-9333996a3e24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.616824] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c1b35f-70c6-4b56-a42f-f882f4f8ec91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.647573] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2ad94c-53e5-4b79-a2e5-c90db9a8ddd3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.655566] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250395ba-a937-4d94-8d02-668172fac4ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.668212] env[61964]: DEBUG nova.compute.provider_tree [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.831020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-10a51687-3a7c-4293-9710-c043e754cd25" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.831149] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 670.831384] env[61964]: DEBUG nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 670.831569] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.848465] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.926884] env[61964]: INFO nova.scheduler.client.report [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Deleted allocations for instance aa526175-3381-48e8-bd5c-73ed37931dbd [ 671.171415] env[61964]: DEBUG nova.scheduler.client.report [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 671.350835] env[61964]: DEBUG nova.network.neutron [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.437555] env[61964]: DEBUG oslo_concurrency.lockutils [None req-61c80fab-acda-4971-8f37-359b88714ba3 tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "aa526175-3381-48e8-bd5c-73ed37931dbd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.231s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.679048] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.679318] env[61964]: ERROR nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Traceback (most recent call last): [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.driver.spawn(context, instance, image_meta, [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] vm_ref = self.build_virtual_machine(instance, [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.679318] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] for vif in network_info: [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self._sync_wrapper(fn, *args, **kwargs) [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.wait() [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self[:] = self._gt.wait() [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self._exit_event.wait() [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] result = hub.switch() [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.679749] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return self.greenlet.switch() [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] result = function(*args, **kwargs) [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] return func(*args, **kwargs) [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise e [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] nwinfo = self.network_api.allocate_for_instance( [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] created_port_ids = self._update_ports_for_instance( [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] with excutils.save_and_reraise_exception(): [ 671.680273] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] self.force_reraise() [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise self.value [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] updated_port = self._update_port( [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] _ensure_no_port_binding_failure(port) [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] raise exception.PortBindingFailed(port_id=port['id']) [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] nova.exception.PortBindingFailed: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. [ 671.680771] env[61964]: ERROR nova.compute.manager [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] [ 671.681159] env[61964]: DEBUG nova.compute.utils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.682199] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.113s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.682934] env[61964]: INFO nova.compute.claims [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.687130] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Build of instance 3a26e916-3d1e-4831-8351-0c12fc775398 was re-scheduled: Binding failed for port 4c812075-d67f-47ee-ae45-c11c0c174ad7, please check neutron logs for more information. {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 671.687130] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Unplugging VIFs for instance {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 671.687130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquiring lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.687130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Acquired lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.687454] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.853493] env[61964]: INFO nova.compute.manager [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 10a51687-3a7c-4293-9710-c043e754cd25] Took 1.02 seconds to deallocate network for instance. [ 671.943102] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 672.207075] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.294601] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.425018] env[61964]: DEBUG nova.compute.manager [None req-12cc33c9-720c-42cb-80f7-0ecef283cd14 tempest-ServerDiagnosticsV248Test-1587390753 tempest-ServerDiagnosticsV248Test-1587390753-project-admin] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 672.426259] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc29a74a-77b4-4d45-90fc-2d10c45ca4ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.434171] env[61964]: INFO nova.compute.manager [None req-12cc33c9-720c-42cb-80f7-0ecef283cd14 tempest-ServerDiagnosticsV248Test-1587390753 tempest-ServerDiagnosticsV248Test-1587390753-project-admin] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Retrieving diagnostics [ 672.435113] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4624a387-02fc-46d7-8a85-3cba98309c3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.483059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.699044] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 672.700075] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9d9294-41e7-4096-9727-14522e4c2d2f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.707370] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 672.707560] env[61964]: ERROR oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk due to incomplete transfer. [ 672.707773] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cb838316-3380-4a4c-8b7d-ed0580b290a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.720021] env[61964]: DEBUG oslo_vmware.rw_handles [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ecfe82-ab76-cbbc-bac2-93908d6138e9/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 672.720021] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Uploaded image 21f5be2c-223a-4f43-a5bb-446e2118758b to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 672.720021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 672.720021] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3cc1fa5b-cc57-41fb-862e-4de6cc5283e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.726932] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 672.726932] env[61964]: value = "task-1040603" [ 672.726932] env[61964]: _type = "Task" [ 672.726932] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.735417] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040603, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.798103] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Releasing lock "refresh_cache-3a26e916-3d1e-4831-8351-0c12fc775398" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.798371] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61964) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 672.798682] env[61964]: DEBUG nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 672.798864] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.816869] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.883626] env[61964]: INFO nova.scheduler.client.report [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted allocations for instance 10a51687-3a7c-4293-9710-c043e754cd25 [ 673.054625] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9878953c-4e43-4531-9bfa-f8d9e3752af8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.064545] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cc620b-fe2d-4008-a73c-0e16ce6600bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.099288] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74523b19-98be-40c2-9eb3-ed10b7ea9fc3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.107310] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc88d655-34b1-41d3-975c-075503e7437d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.121337] env[61964]: DEBUG nova.compute.provider_tree [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.236573] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040603, 'name': Destroy_Task, 'duration_secs': 0.34107} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.236858] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Destroyed the VM [ 673.237198] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 673.237453] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7433f043-4106-4057-8f0a-180338448664 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.244256] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 673.244256] env[61964]: value = "task-1040604" [ 673.244256] env[61964]: _type = "Task" [ 673.244256] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.251650] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040604, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.319796] env[61964]: DEBUG nova.network.neutron [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.393625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40d24647-eeb4-4ecd-b7d5-b507a573f369 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "10a51687-3a7c-4293-9710-c043e754cd25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.250s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.546092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "06631c59-9969-4763-bfe8-472e8dcf8848" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.546482] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.547012] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "06631c59-9969-4763-bfe8-472e8dcf8848-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.547012] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.547135] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.549761] env[61964]: INFO nova.compute.manager [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Terminating instance [ 673.625740] env[61964]: DEBUG nova.scheduler.client.report [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 673.753986] env[61964]: DEBUG oslo_vmware.api [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040604, 'name': RemoveSnapshot_Task, 'duration_secs': 0.344144} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.754297] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 673.754586] env[61964]: INFO nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Took 12.26 seconds to snapshot the instance on the hypervisor. [ 673.822356] env[61964]: INFO nova.compute.manager [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] [instance: 3a26e916-3d1e-4831-8351-0c12fc775398] Took 1.02 seconds to deallocate network for instance. [ 673.895955] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 674.062432] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "refresh_cache-06631c59-9969-4763-bfe8-472e8dcf8848" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.062432] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquired lock "refresh_cache-06631c59-9969-4763-bfe8-472e8dcf8848" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.062432] env[61964]: DEBUG nova.network.neutron [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.132354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.134017] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 674.139015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "0f1ed823-1531-4505-a68a-cb172f804655" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.139015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.139015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "0f1ed823-1531-4505-a68a-cb172f804655-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.139015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.139231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.139231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.503s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.139879] env[61964]: INFO nova.compute.claims [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.142620] env[61964]: INFO nova.compute.manager [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Terminating instance [ 674.259091] env[61964]: DEBUG nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance disappeared during snapshot {{(pid=61964) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 674.271388] env[61964]: DEBUG nova.compute.manager [None req-dea32e19-302c-4944-9a45-3b3e47189d5b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image not found during clean up 21f5be2c-223a-4f43-a5bb-446e2118758b {{(pid=61964) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 674.421680] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.578064] env[61964]: DEBUG nova.network.neutron [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.633211] env[61964]: DEBUG nova.network.neutron [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.639901] env[61964]: DEBUG nova.compute.utils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.641191] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 674.641361] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.653019] env[61964]: DEBUG nova.compute.manager [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 674.653019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 674.653019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c91a365-33d7-4449-9703-9a7ff8115820 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.661167] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 674.661167] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5870f16e-c690-43b2-8350-f2c1b733d58f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.687895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.688154] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.701906] env[61964]: DEBUG nova.policy [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '599501bcaf874b45a3faa4b4ccfb11e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b772e2152b674da8bdc2cb4f726f6772', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.805120] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 674.805120] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 674.805285] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore1] 0f1ed823-1531-4505-a68a-cb172f804655 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 674.805543] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-687e6cdf-6c31-4aa1-bb09-e16ab6f52583 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.814487] env[61964]: DEBUG oslo_vmware.api [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 674.814487] env[61964]: value = "task-1040606" [ 674.814487] env[61964]: _type = "Task" [ 674.814487] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.822625] env[61964]: DEBUG oslo_vmware.api [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.861667] env[61964]: INFO nova.scheduler.client.report [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Deleted allocations for instance 3a26e916-3d1e-4831-8351-0c12fc775398 [ 675.112553] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Successfully created port: edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.135577] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Releasing lock "refresh_cache-06631c59-9969-4763-bfe8-472e8dcf8848" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.135936] env[61964]: DEBUG nova.compute.manager [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 675.136135] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 675.137034] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1c2669-aef5-4e84-9253-a6c3a65ccbda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.145061] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 675.149450] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 675.149450] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc01336f-e52f-41c3-9e35-694113ae92ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.157082] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 675.157082] env[61964]: value = "task-1040607" [ 675.157082] env[61964]: _type = "Task" [ 675.157082] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.166318] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.325038] env[61964]: DEBUG oslo_vmware.api [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040606, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138464} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.325244] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.325436] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 675.325618] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 675.325809] env[61964]: INFO nova.compute.manager [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Took 0.68 seconds to destroy the instance on the hypervisor. [ 675.326068] env[61964]: DEBUG oslo.service.loopingcall [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.326263] env[61964]: DEBUG nova.compute.manager [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 675.326352] env[61964]: DEBUG nova.network.neutron [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.373501] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f8ff8b80-7621-4777-a713-8481d1e0d03a tempest-ServerRescueNegativeTestJSON-583970084 tempest-ServerRescueNegativeTestJSON-583970084-project-member] Lock "3a26e916-3d1e-4831-8351-0c12fc775398" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.570s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.579118] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331918ce-9859-4f2b-945a-550f25397c94 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.583634] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bbad0a-abc9-48b0-8b0a-11610ec0e9bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.614239] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66a2476-9575-436b-9cac-5b7bd6e0cac7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.622023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ec7039-905a-43d8-a098-9bb46871f8df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.637088] env[61964]: DEBUG nova.compute.provider_tree [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.674101] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040607, 'name': PowerOffVM_Task, 'duration_secs': 0.118509} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.674887] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 675.675302] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 675.675979] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4cd04839-3234-44cc-90ba-efd69f1c311d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.705718] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 675.706132] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 675.706468] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Deleting the datastore file [datastore2] 06631c59-9969-4763-bfe8-472e8dcf8848 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.706859] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-daafed2f-247d-4a90-aad2-2a54bfab453d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.715537] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for the task: (returnval){ [ 675.715537] env[61964]: value = "task-1040609" [ 675.715537] env[61964]: _type = "Task" [ 675.715537] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.727337] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.877565] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 675.949705] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.949948] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.950993] env[61964]: DEBUG nova.compute.manager [req-a002a054-ebd3-4a17-b58a-ffcd35e68469 req-d0ef12b2-b5ae-4589-a1da-bd0f833c35c6 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Received event network-vif-deleted-b85285af-4413-4095-940a-792b6a5377bc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 675.951288] env[61964]: INFO nova.compute.manager [req-a002a054-ebd3-4a17-b58a-ffcd35e68469 req-d0ef12b2-b5ae-4589-a1da-bd0f833c35c6 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Neutron deleted interface b85285af-4413-4095-940a-792b6a5377bc; detaching it from the instance and deleting it from the info cache [ 675.951504] env[61964]: DEBUG nova.network.neutron [req-a002a054-ebd3-4a17-b58a-ffcd35e68469 req-d0ef12b2-b5ae-4589-a1da-bd0f833c35c6 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.139569] env[61964]: DEBUG nova.scheduler.client.report [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.154023] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 676.175157] env[61964]: DEBUG nova.network.neutron [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.183528] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 676.183932] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.183932] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 676.184155] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.184486] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 676.184486] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 676.184676] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 676.184898] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 676.184990] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 676.185173] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 676.185350] env[61964]: DEBUG nova.virt.hardware [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 676.186244] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27a224d-b8f6-404d-ad38-49a62d761b7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.195010] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f358d1e-4c55-4be7-9972-4ea112752ea1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.224837] env[61964]: DEBUG oslo_vmware.api [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Task: {'id': task-1040609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088958} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.225090] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.225278] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 676.225459] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 676.225633] env[61964]: INFO nova.compute.manager [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Took 1.09 seconds to destroy the instance on the hypervisor. [ 676.225872] env[61964]: DEBUG oslo.service.loopingcall [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.226324] env[61964]: DEBUG nova.compute.manager [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 676.226426] env[61964]: DEBUG nova.network.neutron [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.399163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.427531] env[61964]: DEBUG nova.network.neutron [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.456816] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7024e0fe-580e-44c2-b9b9-ca9133c3c36c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.465927] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e95c06-bcb2-4708-bc10-94c437b339aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.489197] env[61964]: DEBUG nova.compute.manager [req-a002a054-ebd3-4a17-b58a-ffcd35e68469 req-d0ef12b2-b5ae-4589-a1da-bd0f833c35c6 service nova] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Detach interface failed, port_id=b85285af-4413-4095-940a-792b6a5377bc, reason: Instance 0f1ed823-1531-4505-a68a-cb172f804655 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 676.644559] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.645124] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 676.647838] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.522s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.649127] env[61964]: INFO nova.compute.claims [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.678173] env[61964]: INFO nova.compute.manager [-] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Took 1.35 seconds to deallocate network for instance. [ 676.855096] env[61964]: DEBUG nova.compute.manager [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 676.855384] env[61964]: DEBUG oslo_concurrency.lockutils [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.855608] env[61964]: DEBUG oslo_concurrency.lockutils [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.855784] env[61964]: DEBUG oslo_concurrency.lockutils [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.855958] env[61964]: DEBUG nova.compute.manager [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] No waiting events found dispatching network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 676.856145] env[61964]: WARNING nova.compute.manager [req-447b7f88-ca3d-4c06-87ee-fb1457ce0353 req-8562f7df-25ab-4507-becc-181429393b3b service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received unexpected event network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b for instance with vm_state building and task_state spawning. [ 676.931045] env[61964]: DEBUG nova.network.neutron [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.056135] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Successfully updated port: edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 677.154039] env[61964]: DEBUG nova.compute.utils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.157975] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 677.160114] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 677.184648] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.216378] env[61964]: DEBUG nova.policy [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4bf81d972234720b707312887c1b74f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f864c21d43647f194c05c358e2dc66c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 677.433601] env[61964]: INFO nova.compute.manager [-] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Took 1.21 seconds to deallocate network for instance. [ 677.533109] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Successfully created port: 822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 677.562729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.562878] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.563088] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.666377] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 677.943174] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.114164] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.149674] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09128256-1bf6-4fb1-977f-5d5fad304d52 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.161356] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4acbe54-a402-4e7c-8a98-0907c2c7cd4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.200836] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2631ddb9-4689-4b95-b894-a12bd6cada0e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.209446] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1747bd90-1134-47d2-bf46-4684e75a408c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.223701] env[61964]: DEBUG nova.compute.provider_tree [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.328190] env[61964]: DEBUG nova.network.neutron [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.702301] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 678.727557] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 678.727827] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 678.728098] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 678.728321] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 678.728475] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 678.728630] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 678.728841] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 678.729014] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 678.729307] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 678.729483] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 678.729657] env[61964]: DEBUG nova.virt.hardware [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 678.730683] env[61964]: DEBUG nova.scheduler.client.report [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.734296] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d653086d-2972-41bc-ba0f-281d8e34f164 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.742572] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd08b3dd-12ba-4bf1-a90d-3ee9e292bb51 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.830408] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.830783] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance network_info: |[{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 678.832048] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:0b:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02092ea4-bae0-4e42-b0ab-abc365b4395a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'edda620f-1259-4fbb-afa6-48aef4eda40b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.839169] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating folder: Project (b772e2152b674da8bdc2cb4f726f6772). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 678.839169] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa38d8a1-dfbd-4e04-912e-0d0c1801c02a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.851638] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created folder: Project (b772e2152b674da8bdc2cb4f726f6772) in parent group-v230360. [ 678.851638] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating folder: Instances. Parent ref: group-v230388. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 678.851638] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f43a7627-5e21-4905-a1bf-f26b0a40c526 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.864501] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created folder: Instances in parent group-v230388. [ 678.864685] env[61964]: DEBUG oslo.service.loopingcall [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.865274] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 678.865274] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cbdc58b-50e4-4b7a-8d3d-deccc60511cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.883732] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.883732] env[61964]: value = "task-1040612" [ 678.883732] env[61964]: _type = "Task" [ 678.883732] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.891257] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040612, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.893522] env[61964]: DEBUG nova.compute.manager [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 678.893610] env[61964]: DEBUG nova.compute.manager [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing instance network info cache due to event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 678.893762] env[61964]: DEBUG oslo_concurrency.lockutils [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.893909] env[61964]: DEBUG oslo_concurrency.lockutils [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.894116] env[61964]: DEBUG nova.network.neutron [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 679.196764] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Successfully updated port: 822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 679.237836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.238394] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 679.240924] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.038s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.242421] env[61964]: INFO nova.compute.claims [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.393810] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040612, 'name': CreateVM_Task, 'duration_secs': 0.322708} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.393994] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 679.395304] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.395304] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.395566] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 679.399605] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5ccd275-255d-402b-91ad-999795647405 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.404545] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 679.404545] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523c6b46-7607-367a-32b1-ceb44a12a2b5" [ 679.404545] env[61964]: _type = "Task" [ 679.404545] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.410126] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523c6b46-7607-367a-32b1-ceb44a12a2b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.606219] env[61964]: DEBUG nova.network.neutron [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updated VIF entry in instance network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 679.606591] env[61964]: DEBUG nova.network.neutron [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.699611] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.699727] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.699945] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.747222] env[61964]: DEBUG nova.compute.utils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.750649] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 679.750822] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 679.806441] env[61964]: DEBUG nova.policy [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb7cc75be9a44e2187090e490b8056e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833e91665288466a9ac2141d516ab922', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.913212] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523c6b46-7607-367a-32b1-ceb44a12a2b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009411} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.913526] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.913763] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.914015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.914172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.914354] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.914612] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fd7bc1c-74e8-4869-bded-6142d150389e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.923381] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.923619] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 679.924428] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-783425a3-a71e-4849-aeea-00d87ed7e175 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.930304] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 679.930304] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5258f7ad-08fb-19c7-2265-a187c184e217" [ 679.930304] env[61964]: _type = "Task" [ 679.930304] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.938257] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5258f7ad-08fb-19c7-2265-a187c184e217, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.109788] env[61964]: DEBUG oslo_concurrency.lockutils [req-af7981b4-ef6b-460f-ac50-7dc40a0df5f0 req-52635788-5962-4902-9ed0-c66c866c1717 service nova] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.255216] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 680.259037] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.340473] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Successfully created port: 03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.444026] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5258f7ad-08fb-19c7-2265-a187c184e217, 'name': SearchDatastore_Task, 'duration_secs': 0.008693} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.444589] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56e8f346-4255-4f9d-a738-03f66382d8fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.450748] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 680.450748] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528dc35f-12f2-faa3-2e4a-8254b190cad0" [ 680.450748] env[61964]: _type = "Task" [ 680.450748] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.451712] env[61964]: DEBUG nova.network.neutron [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating instance_info_cache with network_info: [{"id": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "address": "fa:16:3e:64:d3:d6", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap822d4088-78", "ovs_interfaceid": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.464380] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528dc35f-12f2-faa3-2e4a-8254b190cad0, 'name': SearchDatastore_Task, 'duration_secs': 0.009299} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.464516] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.464760] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 680.465030] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d59a06b-4d68-4662-869d-1202503c813e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.472790] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 680.472790] env[61964]: value = "task-1040613" [ 680.472790] env[61964]: _type = "Task" [ 680.472790] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.482545] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.745691] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3366a82b-5f37-4d1c-bd3f-80a986d4eeda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.753394] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85a55f1-2422-4ff4-b902-19ef67b05f66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.787228] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3f20c9-b8ab-4919-a3f8-ca076559f2b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.795504] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d7630b-9f4b-4dc7-a0dd-d485b076e720 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.811089] env[61964]: DEBUG nova.compute.provider_tree [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.959625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.960180] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Instance network_info: |[{"id": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "address": "fa:16:3e:64:d3:d6", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap822d4088-78", "ovs_interfaceid": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 680.960451] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:d3:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bab6a6c3-1c5c-4776-b21b-dec21196d702', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '822d4088-7886-4e3d-8716-59e4a3a18ab0', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.968110] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Creating folder: Project (4f864c21d43647f194c05c358e2dc66c). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.969315] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac0fe713-3c18-493e-8f4d-61194611e31a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.972014] env[61964]: DEBUG nova.compute.manager [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Received event network-vif-plugged-822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 680.972226] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.972434] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.972602] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.972771] env[61964]: DEBUG nova.compute.manager [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] No waiting events found dispatching network-vif-plugged-822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 680.972939] env[61964]: WARNING nova.compute.manager [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Received unexpected event network-vif-plugged-822d4088-7886-4e3d-8716-59e4a3a18ab0 for instance with vm_state building and task_state spawning. [ 680.973149] env[61964]: DEBUG nova.compute.manager [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Received event network-changed-822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 680.973318] env[61964]: DEBUG nova.compute.manager [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Refreshing instance network info cache due to event network-changed-822d4088-7886-4e3d-8716-59e4a3a18ab0. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 680.973507] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Acquiring lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.973646] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Acquired lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.973800] env[61964]: DEBUG nova.network.neutron [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Refreshing network info cache for port 822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.984714] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040613, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.986522] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Created folder: Project (4f864c21d43647f194c05c358e2dc66c) in parent group-v230360. [ 680.986713] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Creating folder: Instances. Parent ref: group-v230391. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.986946] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b832879-1427-40f4-ab21-a6ae1b70414c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.996779] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Created folder: Instances in parent group-v230391. [ 680.996985] env[61964]: DEBUG oslo.service.loopingcall [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.997227] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 680.997396] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f13f1af-a447-4bfb-9307-396bd22bef0e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.017282] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 681.017282] env[61964]: value = "task-1040616" [ 681.017282] env[61964]: _type = "Task" [ 681.017282] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.024733] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040616, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.292531] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 681.314619] env[61964]: DEBUG nova.scheduler.client.report [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 681.320495] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:21:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='4d7c58f5-7642-43b5-a3cb-6aff98a75b9e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-994792262',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 681.321067] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.321334] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 681.321547] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.321893] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 681.322157] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 681.323080] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 681.323080] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 681.323080] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 681.323080] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 681.323284] env[61964]: DEBUG nova.virt.hardware [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 681.324124] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb66212c-e55c-4e56-af93-a8a12a36c8fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.333224] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3c788a-027a-4e8d-af51-76684a49f990 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.486572] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040613, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514505} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.488441] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 681.488441] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.488441] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2b33470-973f-475c-b66a-0af5967e04a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.493849] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 681.493849] env[61964]: value = "task-1040617" [ 681.493849] env[61964]: _type = "Task" [ 681.493849] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.501936] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.527547] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040616, 'name': CreateVM_Task, 'duration_secs': 0.273902} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.527711] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 681.528511] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.528803] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.529147] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 681.529473] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c11caa34-0bf3-4006-b27c-97388abe3375 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.534274] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 681.534274] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5255a5e7-15e7-97b1-a473-dbd0a34b2244" [ 681.534274] env[61964]: _type = "Task" [ 681.534274] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.543212] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5255a5e7-15e7-97b1-a473-dbd0a34b2244, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.750140] env[61964]: DEBUG nova.network.neutron [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updated VIF entry in instance network info cache for port 822d4088-7886-4e3d-8716-59e4a3a18ab0. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 681.750140] env[61964]: DEBUG nova.network.neutron [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating instance_info_cache with network_info: [{"id": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "address": "fa:16:3e:64:d3:d6", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap822d4088-78", "ovs_interfaceid": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.830964] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.831403] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 681.834699] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.578s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.836140] env[61964]: INFO nova.compute.claims [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.003912] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064035} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.004241] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.005015] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213b5567-6987-4740-8353-3299e8537849 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.029019] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.029332] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3ba8e1d-f6b7-4b59-8390-59ff697b9f7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.054058] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5255a5e7-15e7-97b1-a473-dbd0a34b2244, 'name': SearchDatastore_Task, 'duration_secs': 0.0089} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.054908] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Successfully updated port: 03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 682.057516] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.057516] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 682.057594] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.057719] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.058112] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 682.058322] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 682.058322] env[61964]: value = "task-1040618" [ 682.058322] env[61964]: _type = "Task" [ 682.058322] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.058711] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db052490-cc01-4b66-903e-7aaad4899869 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.071129] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.072090] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 682.072282] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 682.073880] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9afc2e1-3394-4bc1-bd81-caa10e3be4fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.080182] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 682.080182] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbbdab-4e36-1913-2c46-bff7bc4562c3" [ 682.080182] env[61964]: _type = "Task" [ 682.080182] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.087621] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbbdab-4e36-1913-2c46-bff7bc4562c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.252801] env[61964]: DEBUG oslo_concurrency.lockutils [req-9577c05f-d7d7-48e5-b579-e519989125ff req-65f37f5f-866c-495a-8e0f-16ad7f636486 service nova] Releasing lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.345173] env[61964]: DEBUG nova.compute.utils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.348664] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 682.348664] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 682.390411] env[61964]: DEBUG nova.policy [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95d7cdb2b34446b6add4f05198ce1110', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbda7c7693524ddb8290f5f719f4067c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.563019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.563019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.563019] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.576173] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040618, 'name': ReconfigVM_Task, 'duration_secs': 0.473156} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.576173] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 682.579141] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c737297-2f55-4198-ad21-7da85149d86e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.591897] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbbdab-4e36-1913-2c46-bff7bc4562c3, 'name': SearchDatastore_Task, 'duration_secs': 0.009031} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.593861] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 682.593861] env[61964]: value = "task-1040619" [ 682.593861] env[61964]: _type = "Task" [ 682.593861] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.595154] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-921e2b6c-00a4-44b1-b821-376f6be3db88 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.607651] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 682.607651] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d51238-1f20-9ea2-1c83-5fa0fc8dddc1" [ 682.607651] env[61964]: _type = "Task" [ 682.607651] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.608772] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040619, 'name': Rename_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.618923] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d51238-1f20-9ea2-1c83-5fa0fc8dddc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008952} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.619071] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.619629] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] a995ff8c-75b9-4089-ad4c-9e6baa33fe8c/a995ff8c-75b9-4089-ad4c-9e6baa33fe8c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 682.619770] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d28edf45-a50e-41f6-b265-943f6c761f27 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.625887] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 682.625887] env[61964]: value = "task-1040620" [ 682.625887] env[61964]: _type = "Task" [ 682.625887] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.634276] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.814747] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Successfully created port: 7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.850132] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 683.110097] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040619, 'name': Rename_Task, 'duration_secs': 0.14379} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.110097] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 683.110314] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c8b38ad-d1f7-4ce4-a638-4d074eec39c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.119938] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 683.119938] env[61964]: value = "task-1040621" [ 683.119938] env[61964]: _type = "Task" [ 683.119938] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.124638] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.137124] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040621, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.139295] env[61964]: DEBUG nova.compute.manager [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Received event network-vif-plugged-03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 683.139505] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Acquiring lock "65566c20-6d69-471c-b098-3c30c01d9955-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.139716] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.139885] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.140457] env[61964]: DEBUG nova.compute.manager [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] No waiting events found dispatching network-vif-plugged-03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 683.140457] env[61964]: WARNING nova.compute.manager [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Received unexpected event network-vif-plugged-03f523b6-4fc2-4fe2-a751-d33c00f6849f for instance with vm_state building and task_state spawning. [ 683.140457] env[61964]: DEBUG nova.compute.manager [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Received event network-changed-03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 683.140604] env[61964]: DEBUG nova.compute.manager [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Refreshing instance network info cache due to event network-changed-03f523b6-4fc2-4fe2-a751-d33c00f6849f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 683.140731] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.146236] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040620, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.295564] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0a14da-6fec-404c-9928-79ce7def9f73 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.303213] env[61964]: DEBUG nova.network.neutron [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.309271] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c34ca16-5b43-49f5-b36b-e30fc5bb0d77 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.344155] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e51a88-3b5b-4bd6-9a06-e38a02aacb6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.352768] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b7d693-6998-4f62-81a1-174096bb3b88 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.359763] env[61964]: INFO nova.virt.block_device [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Booting with volume c397e4a2-9c55-49e6-9341-0fa2a060c38f at /dev/sda [ 683.370433] env[61964]: DEBUG nova.compute.provider_tree [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.414279] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3e4ca52-d108-4e26-88ce-91e688694c4e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.422844] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2417bc6a-7dd6-47b9-8c01-610a721e830c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.445600] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14c1fecf-1d53-436f-950f-51b67c095f20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.454832] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386224c9-eb0e-4be2-acd8-cc1ccbe4ae21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.479722] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23354214-4d8a-4855-981f-fc7d19842eef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.486322] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ee59c5-e2b3-4dba-bdb0-9aec21bc9ca2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.499411] env[61964]: DEBUG nova.virt.block_device [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updating existing volume attachment record: 7feba077-818f-4d19-ad6f-27197acfcbaf {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 683.630616] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040621, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.638870] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040620, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582031} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.639319] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] a995ff8c-75b9-4089-ad4c-9e6baa33fe8c/a995ff8c-75b9-4089-ad4c-9e6baa33fe8c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 683.639698] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 683.639880] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aae5d8a4-8340-4b0b-9127-7cd3c58113c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.646938] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 683.646938] env[61964]: value = "task-1040622" [ 683.646938] env[61964]: _type = "Task" [ 683.646938] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.655685] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.806435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.807013] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Instance network_info: |[{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 683.807542] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.807899] env[61964]: DEBUG nova.network.neutron [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Refreshing network info cache for port 03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 683.809620] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:a6:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd3c6b64-aba2-4bdc-a693-3b4dff3ed861', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03f523b6-4fc2-4fe2-a751-d33c00f6849f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 683.819847] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Creating folder: Project (833e91665288466a9ac2141d516ab922). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 683.821016] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c20c93af-5dd3-409e-9569-8046ede56475 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.832009] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Created folder: Project (833e91665288466a9ac2141d516ab922) in parent group-v230360. [ 683.832601] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Creating folder: Instances. Parent ref: group-v230394. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 683.832601] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b9ca0d6-d6fb-4a98-82f9-fd8a633ad9ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.842032] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Created folder: Instances in parent group-v230394. [ 683.842237] env[61964]: DEBUG oslo.service.loopingcall [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.842431] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 683.842629] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bacf45ab-184e-4612-a1e3-7bf0889b3f74 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.861905] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 683.861905] env[61964]: value = "task-1040625" [ 683.861905] env[61964]: _type = "Task" [ 683.861905] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.869967] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040625, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.872929] env[61964]: DEBUG nova.scheduler.client.report [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 684.132324] env[61964]: DEBUG oslo_vmware.api [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040621, 'name': PowerOnVM_Task, 'duration_secs': 0.586316} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.132695] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 684.132867] env[61964]: INFO nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Took 7.98 seconds to spawn the instance on the hypervisor. [ 684.133194] env[61964]: DEBUG nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 684.133911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44ec474-fb1d-4277-a484-88623b4f770e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.156121] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065536} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.156400] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.157167] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075a7540-57a2-4694-8a37-be891c48b603 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.182085] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] a995ff8c-75b9-4089-ad4c-9e6baa33fe8c/a995ff8c-75b9-4089-ad4c-9e6baa33fe8c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.182960] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78e19d3d-fb8d-467d-93ca-c7c0bb99b8c2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.202907] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 684.202907] env[61964]: value = "task-1040626" [ 684.202907] env[61964]: _type = "Task" [ 684.202907] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.211353] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040626, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.373412] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040625, 'name': CreateVM_Task, 'duration_secs': 0.371726} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.373738] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 684.375386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.375386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.375386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 684.375386] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aacc3563-75cf-4eac-bf8b-763ad75d3598 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.377699] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.378238] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 684.380901] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.487s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.381632] env[61964]: DEBUG nova.objects.instance [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lazy-loading 'resources' on Instance uuid 9f216f01-dc9f-459f-88be-3a5439d3701a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 684.388231] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 684.388231] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e0dcf1-8ae3-5125-dc5f-baccd99d3ca4" [ 684.388231] env[61964]: _type = "Task" [ 684.388231] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.396532] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e0dcf1-8ae3-5125-dc5f-baccd99d3ca4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.501506] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Successfully updated port: 7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 684.567789] env[61964]: DEBUG nova.network.neutron [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updated VIF entry in instance network info cache for port 03f523b6-4fc2-4fe2-a751-d33c00f6849f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 684.568228] env[61964]: DEBUG nova.network.neutron [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.651874] env[61964]: INFO nova.compute.manager [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Took 42.14 seconds to build instance. [ 684.713024] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040626, 'name': ReconfigVM_Task, 'duration_secs': 0.381533} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.713024] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfigured VM instance instance-00000021 to attach disk [datastore2] a995ff8c-75b9-4089-ad4c-9e6baa33fe8c/a995ff8c-75b9-4089-ad4c-9e6baa33fe8c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 684.713511] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-832d727e-fca7-4c55-822a-abef6d3cf451 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.720090] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 684.720090] env[61964]: value = "task-1040627" [ 684.720090] env[61964]: _type = "Task" [ 684.720090] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.728859] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040627, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.884551] env[61964]: DEBUG nova.compute.utils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.889762] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 684.889762] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 684.902270] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e0dcf1-8ae3-5125-dc5f-baccd99d3ca4, 'name': SearchDatastore_Task, 'duration_secs': 0.010287} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.902673] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.902933] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 684.903548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.903548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.903692] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 684.903858] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-734ba037-fb05-40e0-9085-d52c26959cc7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.912273] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 684.913262] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 684.916527] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-701e7584-a2a5-4c02-ab59-6a2868d5a736 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.923386] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 684.923386] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528da34b-36fd-98fb-9994-13cdda58623c" [ 684.923386] env[61964]: _type = "Task" [ 684.923386] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.932713] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528da34b-36fd-98fb-9994-13cdda58623c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.947479] env[61964]: DEBUG nova.policy [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dbb0a78b0154ebcb7e5d54b5daecc8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a474c9bb101143999bb79396fe1c52de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.005680] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.005680] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquired lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.005892] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.073336] env[61964]: DEBUG oslo_concurrency.lockutils [req-9579145b-af89-429e-a818-1da2019f1c39 req-847af50c-eef3-4dce-a0b2-583940c1362a service nova] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.154688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a552b257-22d7-481f-993a-762c3a26523a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.650s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.230489] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040627, 'name': Rename_Task, 'duration_secs': 0.145842} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.230763] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 685.231013] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81b07713-a19f-4a89-8ce0-fa40a50310d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.239393] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 685.239393] env[61964]: value = "task-1040628" [ 685.239393] env[61964]: _type = "Task" [ 685.239393] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.249939] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.328998] env[61964]: DEBUG nova.compute.manager [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Received event network-vif-plugged-7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 685.329282] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Acquiring lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.329531] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.329723] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.329921] env[61964]: DEBUG nova.compute.manager [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] No waiting events found dispatching network-vif-plugged-7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 685.330149] env[61964]: WARNING nova.compute.manager [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Received unexpected event network-vif-plugged-7210f188-7099-4e52-ad5f-f20da7bc5d91 for instance with vm_state building and task_state spawning. [ 685.330346] env[61964]: DEBUG nova.compute.manager [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Received event network-changed-7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 685.330532] env[61964]: DEBUG nova.compute.manager [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Refreshing instance network info cache due to event network-changed-7210f188-7099-4e52-ad5f-f20da7bc5d91. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 685.330736] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Acquiring lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.342035] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce25dd34-3c4a-4de9-ad6c-b46dce599e14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.350668] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2bae77-ac9e-4cb6-9cfe-b0b68b003500 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.355027] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Successfully created port: f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.386867] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d73cc68-c2b9-4dbb-b81f-2804f16623c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.390530] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 685.398986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd50ba3-1ecd-4dfe-b8c6-bfc26da03ccc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.412197] env[61964]: DEBUG nova.compute.provider_tree [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.438853] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528da34b-36fd-98fb-9994-13cdda58623c, 'name': SearchDatastore_Task, 'duration_secs': 0.008383} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.439736] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62c8caa7-fe7e-4d13-b8a9-507debf5bd69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.445505] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 685.445505] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521bbfcf-bcd1-2e3c-adbe-2073459d9038" [ 685.445505] env[61964]: _type = "Task" [ 685.445505] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.453820] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521bbfcf-bcd1-2e3c-adbe-2073459d9038, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.543556] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.604632] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 685.605259] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 685.605489] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.605673] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 685.605865] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.606027] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 685.606245] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 685.606590] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 685.606776] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 685.606995] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 685.607267] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 685.607576] env[61964]: DEBUG nova.virt.hardware [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 685.608684] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63a843c-6909-4bde-89b3-343b8d9fe97f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.619268] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089931e4-d282-4883-aa5b-87345eeda20e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.657694] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 685.728261] env[61964]: DEBUG nova.network.neutron [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updating instance_info_cache with network_info: [{"id": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "address": "fa:16:3e:fd:be:5f", "network": {"id": "af0be9de-9c19-46e4-a4de-6a5ae7c7fb44", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2115936393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbda7c7693524ddb8290f5f719f4067c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7210f188-70", "ovs_interfaceid": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.750350] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040628, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.916103] env[61964]: DEBUG nova.scheduler.client.report [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.956175] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521bbfcf-bcd1-2e3c-adbe-2073459d9038, 'name': SearchDatastore_Task, 'duration_secs': 0.014044} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.956449] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.956761] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 685.957037] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31874ac2-8976-42e8-827a-10acfd94e542 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.964343] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 685.964343] env[61964]: value = "task-1040629" [ 685.964343] env[61964]: _type = "Task" [ 685.964343] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.973434] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.030309] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "595c94f1-9dec-454d-b301-5656287f53de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.030582] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.181486] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.231697] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Releasing lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.232058] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance network_info: |[{"id": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "address": "fa:16:3e:fd:be:5f", "network": {"id": "af0be9de-9c19-46e4-a4de-6a5ae7c7fb44", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2115936393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbda7c7693524ddb8290f5f719f4067c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7210f188-70", "ovs_interfaceid": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 686.232400] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Acquired lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.232592] env[61964]: DEBUG nova.network.neutron [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Refreshing network info cache for port 7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 686.233888] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:be:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eb2eaecd-9701-4504-9fcb-fb1a420ead72', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7210f188-7099-4e52-ad5f-f20da7bc5d91', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 686.243593] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Creating folder: Project (cbda7c7693524ddb8290f5f719f4067c). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.244909] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efa790f8-7803-4ff1-8cb5-3aba7d9d7330 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.256537] env[61964]: DEBUG oslo_vmware.api [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040628, 'name': PowerOnVM_Task, 'duration_secs': 0.574889} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.256805] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 686.257015] env[61964]: INFO nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Took 7.55 seconds to spawn the instance on the hypervisor. [ 686.257212] env[61964]: DEBUG nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 686.258344] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fee40c-5c94-4402-8dd5-eaa31773da43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.262385] env[61964]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 686.262479] env[61964]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61964) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 686.262778] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Folder already exists: Project (cbda7c7693524ddb8290f5f719f4067c). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 686.262975] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Creating folder: Instances. Parent ref: group-v230366. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 686.263634] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bd50e87-d036-4f6f-80cd-4fef8da5a024 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.277888] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Created folder: Instances in parent group-v230366. [ 686.278184] env[61964]: DEBUG oslo.service.loopingcall [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.278395] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 686.278610] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ec6735f-b03d-45fd-b9d1-f6dd1bbe7538 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.307049] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 686.307049] env[61964]: value = "task-1040632" [ 686.307049] env[61964]: _type = "Task" [ 686.307049] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.317204] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040632, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.400238] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 686.420876] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.424917] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.150s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.426485] env[61964]: INFO nova.compute.claims [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.435613] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 686.435897] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.436080] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.436281] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.436433] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.436581] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 686.436797] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 686.436962] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 686.437326] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 686.437326] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 686.437497] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 686.438393] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55c11b2-de62-4dd9-a562-97083fde8245 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.442505] env[61964]: INFO nova.scheduler.client.report [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Deleted allocations for instance 9f216f01-dc9f-459f-88be-3a5439d3701a [ 686.449795] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8e77cc-dcef-42dc-bbdb-398d04a46006 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.474459] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040629, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466519} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.474813] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 686.475109] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 686.475297] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c012a171-c311-4431-8beb-10181e630926 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.481515] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 686.481515] env[61964]: value = "task-1040633" [ 686.481515] env[61964]: _type = "Task" [ 686.481515] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.490983] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040633, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.786770] env[61964]: INFO nova.compute.manager [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Took 35.17 seconds to build instance. [ 686.819924] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040632, 'name': CreateVM_Task, 'duration_secs': 0.3702} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.820109] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 686.820791] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'delete_on_termination': True, 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': '7feba077-818f-4d19-ad6f-27197acfcbaf', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230369', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'name': 'volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cdfc0bf1-d603-4580-8527-8b06e5ae0799', 'attached_at': '', 'detached_at': '', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'serial': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f'}, 'volume_type': None}], 'swap': None} {{(pid=61964) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 686.820993] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Root volume attach. Driver type: vmdk {{(pid=61964) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 686.821867] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019b86fe-1642-4e12-bac1-9c73c0f1c937 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.831197] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be4bb66-1972-47b9-81e7-1cdab1b6ca81 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.843795] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67068518-316c-4bc2-960b-393bdf04a69e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.851534] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-7317c32d-e4bb-4e32-9dec-f6a133209af5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.862265] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 686.862265] env[61964]: value = "task-1040634" [ 686.862265] env[61964]: _type = "Task" [ 686.862265] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.868492] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040634, 'name': RelocateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.936721] env[61964]: DEBUG nova.compute.manager [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Received event network-vif-plugged-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 686.936958] env[61964]: DEBUG oslo_concurrency.lockutils [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] Acquiring lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.937196] env[61964]: DEBUG oslo_concurrency.lockutils [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.937364] env[61964]: DEBUG oslo_concurrency.lockutils [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.937553] env[61964]: DEBUG nova.compute.manager [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] No waiting events found dispatching network-vif-plugged-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 686.937737] env[61964]: WARNING nova.compute.manager [req-6e041922-4836-48e9-ab52-040686385578 req-d40a366f-f2f3-43c9-9f93-9b12f9cb294f service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Received unexpected event network-vif-plugged-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 for instance with vm_state building and task_state spawning. [ 686.958541] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71e243a6-4f80-491e-a249-4b30918acf51 tempest-ServersAaction247Test-1123561841 tempest-ServersAaction247Test-1123561841-project-member] Lock "9f216f01-dc9f-459f-88be-3a5439d3701a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.350s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.995798] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040633, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062047} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.996456] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 686.997366] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47600a6-d718-4120-8dae-15ec97ac7b5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.021713] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 687.022412] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d658bcc-e0dd-4a7a-a8d4-e7cb01afdcf0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.042775] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 687.042775] env[61964]: value = "task-1040635" [ 687.042775] env[61964]: _type = "Task" [ 687.042775] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.049248] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Successfully updated port: f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 687.050460] env[61964]: DEBUG nova.network.neutron [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updated VIF entry in instance network info cache for port 7210f188-7099-4e52-ad5f-f20da7bc5d91. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 687.050887] env[61964]: DEBUG nova.network.neutron [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updating instance_info_cache with network_info: [{"id": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "address": "fa:16:3e:fd:be:5f", "network": {"id": "af0be9de-9c19-46e4-a4de-6a5ae7c7fb44", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2115936393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbda7c7693524ddb8290f5f719f4067c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7210f188-70", "ovs_interfaceid": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.057777] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040635, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.288801] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2c38f2-f12a-4dde-9778-29b606dd4b86 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.779s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.371876] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040634, 'name': RelocateVM_Task} progress is 20%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.556415] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.556735] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.556921] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.561809] env[61964]: DEBUG oslo_concurrency.lockutils [req-e832caf4-1ed1-4632-9d0a-6dffc89d55c1 req-b8bce0f5-e76c-4614-84ba-9f14f57adab1 service nova] Releasing lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.561809] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040635, 'name': ReconfigVM_Task, 'duration_secs': 0.274665} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.565352] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 687.566469] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4614224e-9875-4363-aa94-77c0cc46facf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.574133] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 687.574133] env[61964]: value = "task-1040636" [ 687.574133] env[61964]: _type = "Task" [ 687.574133] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.586860] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040636, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.791850] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 687.858283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d251516-db1e-433a-8ed0-d32c88d83df0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.866895] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f42cba7-de27-4a1f-8944-df4737a5310c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.874078] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040634, 'name': RelocateVM_Task, 'duration_secs': 0.824663} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.874656] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 687.874874] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230369', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'name': 'volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cdfc0bf1-d603-4580-8527-8b06e5ae0799', 'attached_at': '', 'detached_at': '', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'serial': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 687.875749] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5981a9-9a31-40bc-9e33-5b4387cc205e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.910194] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecf1f64-28ab-4e77-9efa-9216a8bdf5be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.923601] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4daee45b-9aaa-4035-84d9-2846d4ed7550 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.929019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0dfaeb-8d3b-4c84-ab33-377e9923ec61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.951312] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f/volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 687.952050] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aeb292e7-b04b-4d7c-b8fb-3e421061900d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.978661] env[61964]: DEBUG nova.compute.provider_tree [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.986119] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 687.986119] env[61964]: value = "task-1040637" [ 687.986119] env[61964]: _type = "Task" [ 687.986119] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.999164] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040637, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.085152] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040636, 'name': Rename_Task, 'duration_secs': 0.136383} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.085486] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 688.085746] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d41541f-b55f-4bc5-8d35-9aa2d600dd3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.091783] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 688.091783] env[61964]: value = "task-1040638" [ 688.091783] env[61964]: _type = "Task" [ 688.091783] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.102191] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040638, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.129338] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.320909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.418632] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Updating instance_info_cache with network_info: [{"id": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "address": "fa:16:3e:c0:8d:78", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6610a7b-e3", "ovs_interfaceid": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.476218] env[61964]: DEBUG nova.compute.manager [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Received event network-changed-822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 688.476218] env[61964]: DEBUG nova.compute.manager [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Refreshing instance network info cache due to event network-changed-822d4088-7886-4e3d-8716-59e4a3a18ab0. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 688.476218] env[61964]: DEBUG oslo_concurrency.lockutils [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] Acquiring lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.476218] env[61964]: DEBUG oslo_concurrency.lockutils [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] Acquired lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.476493] env[61964]: DEBUG nova.network.neutron [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Refreshing network info cache for port 822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 688.484158] env[61964]: DEBUG nova.scheduler.client.report [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 688.498341] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040637, 'name': ReconfigVM_Task, 'duration_secs': 0.308106} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.498877] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Reconfigured VM instance instance-00000023 to attach disk [datastore1] volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f/volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 688.504645] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd273f5d-eca0-4d46-98f5-713a3b562841 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.521438] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 688.521438] env[61964]: value = "task-1040639" [ 688.521438] env[61964]: _type = "Task" [ 688.521438] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.532125] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040639, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.605917] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040638, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.927837] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.928199] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Instance network_info: |[{"id": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "address": "fa:16:3e:c0:8d:78", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6610a7b-e3", "ovs_interfaceid": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 688.928686] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:8d:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 688.936448] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating folder: Project (a474c9bb101143999bb79396fe1c52de). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 688.936875] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1213a0e-3abb-4306-a1e6-6c580ea098e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.948891] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created folder: Project (a474c9bb101143999bb79396fe1c52de) in parent group-v230360. [ 688.949117] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating folder: Instances. Parent ref: group-v230399. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 688.949485] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad4bdb04-0995-4c3c-ac62-1d89dca7780e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.959866] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created folder: Instances in parent group-v230399. [ 688.959968] env[61964]: DEBUG oslo.service.loopingcall [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.961028] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 688.961028] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2edb959-64fd-4fae-8f23-09892790656c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.977292] env[61964]: DEBUG nova.compute.manager [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Received event network-changed-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 688.977489] env[61964]: DEBUG nova.compute.manager [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Refreshing instance network info cache due to event network-changed-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 688.977708] env[61964]: DEBUG oslo_concurrency.lockutils [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] Acquiring lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.977859] env[61964]: DEBUG oslo_concurrency.lockutils [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] Acquired lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.978026] env[61964]: DEBUG nova.network.neutron [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Refreshing network info cache for port f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 688.985374] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 688.985374] env[61964]: value = "task-1040642" [ 688.985374] env[61964]: _type = "Task" [ 688.985374] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.986751] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.987309] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 688.993927] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.040s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.995645] env[61964]: INFO nova.compute.claims [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.006020] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040642, 'name': CreateVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.035609] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040639, 'name': ReconfigVM_Task, 'duration_secs': 0.138869} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.036380] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230369', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'name': 'volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cdfc0bf1-d603-4580-8527-8b06e5ae0799', 'attached_at': '', 'detached_at': '', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'serial': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 689.039176] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25b27923-3e27-47c2-b0fe-697df8a000a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.048282] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 689.048282] env[61964]: value = "task-1040643" [ 689.048282] env[61964]: _type = "Task" [ 689.048282] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.058568] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040643, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.103075] env[61964]: DEBUG oslo_vmware.api [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040638, 'name': PowerOnVM_Task, 'duration_secs': 0.514264} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.103396] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 689.103610] env[61964]: INFO nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Took 7.81 seconds to spawn the instance on the hypervisor. [ 689.103799] env[61964]: DEBUG nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 689.104818] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9932ea1-c252-4b3d-86a7-4514bb1ecb55 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.300723] env[61964]: DEBUG nova.network.neutron [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updated VIF entry in instance network info cache for port 822d4088-7886-4e3d-8716-59e4a3a18ab0. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 689.301400] env[61964]: DEBUG nova.network.neutron [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating instance_info_cache with network_info: [{"id": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "address": "fa:16:3e:64:d3:d6", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap822d4088-78", "ovs_interfaceid": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.492689] env[61964]: DEBUG nova.compute.utils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.497823] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 689.497943] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 689.499749] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040642, 'name': CreateVM_Task, 'duration_secs': 0.355229} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.502091] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 689.503149] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.503149] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.503498] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 689.504409] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39f3fcec-d5de-4b9c-83d0-ad82667355f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.510741] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 689.510741] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524af2e6-c54f-e021-e0e7-28b9f8d68fcb" [ 689.510741] env[61964]: _type = "Task" [ 689.510741] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.520246] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524af2e6-c54f-e021-e0e7-28b9f8d68fcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.542655] env[61964]: DEBUG nova.policy [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dbb0a78b0154ebcb7e5d54b5daecc8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a474c9bb101143999bb79396fe1c52de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.557127] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040643, 'name': Rename_Task, 'duration_secs': 0.173504} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.557400] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 689.557634] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b54fc47-a0b8-4ded-81c8-7218d2daf721 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.564064] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 689.564064] env[61964]: value = "task-1040644" [ 689.564064] env[61964]: _type = "Task" [ 689.564064] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.571278] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.625659] env[61964]: INFO nova.compute.manager [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Took 37.53 seconds to build instance. [ 689.724383] env[61964]: DEBUG nova.network.neutron [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Updated VIF entry in instance network info cache for port f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 689.724778] env[61964]: DEBUG nova.network.neutron [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Updating instance_info_cache with network_info: [{"id": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "address": "fa:16:3e:c0:8d:78", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6610a7b-e3", "ovs_interfaceid": "f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.804697] env[61964]: DEBUG oslo_concurrency.lockutils [req-1d24d3ba-0b21-4195-b29d-a3ee3f712631 req-9612ed29-644c-479f-bbfc-db67ed2d0431 service nova] Releasing lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.841513] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Successfully created port: bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.002885] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 690.020949] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524af2e6-c54f-e021-e0e7-28b9f8d68fcb, 'name': SearchDatastore_Task, 'duration_secs': 0.023863} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.021301] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.021687] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.021774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.021923] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.022116] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.022387] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fdbf6f8-1cd5-4868-8865-d3ddeb20eaa0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.030605] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.030895] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 690.031515] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-985ece0e-bc20-49f4-9e90-b6225250bef9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.036574] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 690.036574] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5239cd91-ba32-d026-3b86-c8792cb05573" [ 690.036574] env[61964]: _type = "Task" [ 690.036574] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.046144] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5239cd91-ba32-d026-3b86-c8792cb05573, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.074805] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040644, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.128332] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf27e1de-353d-4e49-b52c-430037892b88 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.939s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.228104] env[61964]: DEBUG oslo_concurrency.lockutils [req-703709ed-df64-4477-bebd-13cc0d53c97b req-71ede38f-569a-427d-a204-9655b69097f9 service nova] Releasing lock "refresh_cache-d99859dc-2206-42d9-ae6d-8294fbd6942c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.457065] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d2f375-1bae-4398-ba5a-9407621848da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.467147] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1c0ada-28f0-4313-adba-f7de45958fa2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.499615] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3ec61c-c2ad-4397-a4ac-38c53b686f5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.508878] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b207ea-135b-41a9-8fa6-761e89c43ba9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.530291] env[61964]: DEBUG nova.compute.provider_tree [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.546856] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5239cd91-ba32-d026-3b86-c8792cb05573, 'name': SearchDatastore_Task, 'duration_secs': 0.009702} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.547672] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fe1eac4-67c8-4ae0-9007-84f3c0f2fbce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.553398] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 690.553398] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bcb48e-5ac1-8ef2-be6f-d0a89fffdf9b" [ 690.553398] env[61964]: _type = "Task" [ 690.553398] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.563331] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bcb48e-5ac1-8ef2-be6f-d0a89fffdf9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.575269] env[61964]: DEBUG oslo_vmware.api [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040644, 'name': PowerOnVM_Task, 'duration_secs': 0.717653} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.575563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 690.575765] env[61964]: INFO nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Took 4.97 seconds to spawn the instance on the hypervisor. [ 690.575945] env[61964]: DEBUG nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 690.576697] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c55ef6d-c378-440c-bc9b-94b6e220d370 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.631062] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 691.023095] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 691.033355] env[61964]: DEBUG nova.scheduler.client.report [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.051332] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 691.051587] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.051750] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.051953] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.052119] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.052273] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 691.052482] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 691.052646] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 691.052818] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 691.052984] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 691.053185] env[61964]: DEBUG nova.virt.hardware [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 691.054427] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049baca3-bd30-4601-93de-ed0f575bed60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.069602] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117bacb6-dc7a-4bd2-a13d-b51784ec7c18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.073827] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bcb48e-5ac1-8ef2-be6f-d0a89fffdf9b, 'name': SearchDatastore_Task, 'duration_secs': 0.009302} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.074365] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.074629] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d99859dc-2206-42d9-ae6d-8294fbd6942c/d99859dc-2206-42d9-ae6d-8294fbd6942c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 691.075222] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0cf2e76-3671-477c-95f6-8ab5867d1b05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.091950] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 691.091950] env[61964]: value = "task-1040645" [ 691.091950] env[61964]: _type = "Task" [ 691.091950] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.096857] env[61964]: INFO nova.compute.manager [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Took 38.92 seconds to build instance. [ 691.103747] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.156176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.420391] env[61964]: DEBUG nova.compute.manager [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Received event network-vif-plugged-bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 691.420391] env[61964]: DEBUG oslo_concurrency.lockutils [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] Acquiring lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.420391] env[61964]: DEBUG oslo_concurrency.lockutils [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.420391] env[61964]: DEBUG oslo_concurrency.lockutils [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.420391] env[61964]: DEBUG nova.compute.manager [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] No waiting events found dispatching network-vif-plugged-bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 691.420617] env[61964]: WARNING nova.compute.manager [req-dd1fc2df-1695-43e1-898a-5f140f78602f req-81986876-f4aa-401c-8b60-f6da5e00d01a service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Received unexpected event network-vif-plugged-bd317133-fb07-49c5-80d0-008894442dc7 for instance with vm_state building and task_state spawning. [ 691.541575] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.542043] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 691.546856] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.336s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.547137] env[61964]: DEBUG nova.objects.instance [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lazy-loading 'resources' on Instance uuid d8d12f21-5af8-4156-8e89-3f719199ca0a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 691.595915] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Successfully updated port: bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 691.604838] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b5eb8080-d664-403d-aca5-21707c3c44f2 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.217s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.612933] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484267} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.613497] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d99859dc-2206-42d9-ae6d-8294fbd6942c/d99859dc-2206-42d9-ae6d-8294fbd6942c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 691.614489] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.614692] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea51bdf9-d990-48d9-8f90-3348c2040843 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.623007] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 691.623007] env[61964]: value = "task-1040646" [ 691.623007] env[61964]: _type = "Task" [ 691.623007] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.632439] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.658183] env[61964]: DEBUG nova.compute.manager [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Received event network-changed-7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 691.658561] env[61964]: DEBUG nova.compute.manager [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Refreshing instance network info cache due to event network-changed-7210f188-7099-4e52-ad5f-f20da7bc5d91. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 691.658952] env[61964]: DEBUG oslo_concurrency.lockutils [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] Acquiring lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.659243] env[61964]: DEBUG oslo_concurrency.lockutils [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] Acquired lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.659560] env[61964]: DEBUG nova.network.neutron [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Refreshing network info cache for port 7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 691.684229] env[61964]: DEBUG nova.compute.manager [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Stashing vm_state: active {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 692.050652] env[61964]: DEBUG nova.compute.utils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.052616] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 692.052858] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 692.106194] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.106361] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.106522] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.107791] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 692.119792] env[61964]: DEBUG nova.policy [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.135430] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.239767} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.135583] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.142855] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c89fcfc-23f7-4986-aa77-cea2063b2145 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.171470] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] d99859dc-2206-42d9-ae6d-8294fbd6942c/d99859dc-2206-42d9-ae6d-8294fbd6942c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.177452] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fc5c7ff-f0d6-4a27-baf4-89c9e489083d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.202411] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 692.202411] env[61964]: value = "task-1040647" [ 692.202411] env[61964]: _type = "Task" [ 692.202411] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.207490] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.213819] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040647, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.437677] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Successfully created port: 74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.488694] env[61964]: DEBUG nova.network.neutron [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updated VIF entry in instance network info cache for port 7210f188-7099-4e52-ad5f-f20da7bc5d91. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 692.489088] env[61964]: DEBUG nova.network.neutron [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updating instance_info_cache with network_info: [{"id": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "address": "fa:16:3e:fd:be:5f", "network": {"id": "af0be9de-9c19-46e4-a4de-6a5ae7c7fb44", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-2115936393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbda7c7693524ddb8290f5f719f4067c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eb2eaecd-9701-4504-9fcb-fb1a420ead72", "external-id": "nsx-vlan-transportzone-433", "segmentation_id": 433, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7210f188-70", "ovs_interfaceid": "7210f188-7099-4e52-ad5f-f20da7bc5d91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.535075] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6dceff-bed1-486c-a2fb-2e356bd72e69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.542838] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6c8d88-53fd-439b-b5bf-4fc2c0b1747e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.575319] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 692.578864] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4daaaad5-253d-4eaf-b983-959a57c64772 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.588193] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44786fd2-1380-4215-9ae2-ea664ad8f8de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.603667] env[61964]: DEBUG nova.compute.provider_tree [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.632351] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.662891] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.715556] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040647, 'name': ReconfigVM_Task, 'duration_secs': 0.351816} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.715556] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Reconfigured VM instance instance-00000024 to attach disk [datastore1] d99859dc-2206-42d9-ae6d-8294fbd6942c/d99859dc-2206-42d9-ae6d-8294fbd6942c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 692.715556] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79098ba5-c52a-40ab-ad51-4761579d6e8c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.725098] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 692.725098] env[61964]: value = "task-1040648" [ 692.725098] env[61964]: _type = "Task" [ 692.725098] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.733678] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040648, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.872578] env[61964]: DEBUG nova.network.neutron [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Updating instance_info_cache with network_info: [{"id": "bd317133-fb07-49c5-80d0-008894442dc7", "address": "fa:16:3e:11:1a:25", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd317133-fb", "ovs_interfaceid": "bd317133-fb07-49c5-80d0-008894442dc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.991937] env[61964]: DEBUG oslo_concurrency.lockutils [req-5b5dfca3-0079-47a1-a014-a7d9e34935a9 req-fe937eb0-598f-475d-a5c3-1305483955b9 service nova] Releasing lock "refresh_cache-cdfc0bf1-d603-4580-8527-8b06e5ae0799" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.110490] env[61964]: DEBUG nova.scheduler.client.report [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 693.236656] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040648, 'name': Rename_Task, 'duration_secs': 0.14375} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.236942] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 693.237216] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da9ffddc-3522-49c9-bf03-4167b9ed1993 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.243860] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 693.243860] env[61964]: value = "task-1040649" [ 693.243860] env[61964]: _type = "Task" [ 693.243860] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.251634] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.376684] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.377044] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Instance network_info: |[{"id": "bd317133-fb07-49c5-80d0-008894442dc7", "address": "fa:16:3e:11:1a:25", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd317133-fb", "ovs_interfaceid": "bd317133-fb07-49c5-80d0-008894442dc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 693.377507] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:1a:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd317133-fb07-49c5-80d0-008894442dc7', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.385254] env[61964]: DEBUG oslo.service.loopingcall [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.385490] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 693.386106] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c1c4703-6f2e-4d3b-8274-73eb99523bd5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.405257] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.405257] env[61964]: value = "task-1040650" [ 693.405257] env[61964]: _type = "Task" [ 693.405257] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.413442] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040650, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.446293] env[61964]: DEBUG nova.compute.manager [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Received event network-changed-bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 693.446293] env[61964]: DEBUG nova.compute.manager [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Refreshing instance network info cache due to event network-changed-bd317133-fb07-49c5-80d0-008894442dc7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 693.446507] env[61964]: DEBUG oslo_concurrency.lockutils [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] Acquiring lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.446709] env[61964]: DEBUG oslo_concurrency.lockutils [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] Acquired lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.448060] env[61964]: DEBUG nova.network.neutron [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Refreshing network info cache for port bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.590051] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 693.613721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.067s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.618085] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 693.618365] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.618534] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 693.618724] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.618949] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 693.619042] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 693.619257] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 693.619423] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 693.619595] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 693.619814] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 693.619924] env[61964]: DEBUG nova.virt.hardware [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 693.620502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.138s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.622070] env[61964]: INFO nova.compute.claims [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.625267] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee4c863-4c19-4ade-aaa7-52cff82c900f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.634714] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9ff22d-20ca-4caf-84b2-07ef3bc22d72 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.650630] env[61964]: INFO nova.scheduler.client.report [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Deleted allocations for instance d8d12f21-5af8-4156-8e89-3f719199ca0a [ 693.753599] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040649, 'name': PowerOnVM_Task, 'duration_secs': 0.508771} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.753984] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 693.754109] env[61964]: INFO nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Took 7.35 seconds to spawn the instance on the hypervisor. [ 693.754301] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 693.755094] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec43b9cb-c772-49fa-b64d-42a084a94918 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.918797] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040650, 'name': CreateVM_Task, 'duration_secs': 0.359045} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.920232] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 693.921141] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.921475] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.922495] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 693.922863] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b90e108-a555-44d2-8721-5f251eb49c08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.928214] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 693.928214] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5253f508-5b8f-071d-c507-fd8a9a84c34f" [ 693.928214] env[61964]: _type = "Task" [ 693.928214] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.942611] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5253f508-5b8f-071d-c507-fd8a9a84c34f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.066949] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Successfully updated port: 74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 694.161451] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b360029d-90cd-4b06-9e9f-16a40c6b787b tempest-ServerRescueTestJSONUnderV235-1354977986 tempest-ServerRescueTestJSONUnderV235-1354977986-project-member] Lock "d8d12f21-5af8-4156-8e89-3f719199ca0a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.398s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.231189] env[61964]: DEBUG nova.network.neutron [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Updated VIF entry in instance network info cache for port bd317133-fb07-49c5-80d0-008894442dc7. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 694.231565] env[61964]: DEBUG nova.network.neutron [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Updating instance_info_cache with network_info: [{"id": "bd317133-fb07-49c5-80d0-008894442dc7", "address": "fa:16:3e:11:1a:25", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd317133-fb", "ovs_interfaceid": "bd317133-fb07-49c5-80d0-008894442dc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.272860] env[61964]: INFO nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Took 40.04 seconds to build instance. [ 694.439088] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5253f508-5b8f-071d-c507-fd8a9a84c34f, 'name': SearchDatastore_Task, 'duration_secs': 0.016938} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.439659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.439802] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 694.440040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.440198] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.440387] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 694.440688] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e03129ea-0fa1-44be-9619-16bb9dd95a89 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.448946] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 694.449161] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 694.449886] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c41662f-fa97-4c6a-a470-e16555bf2e8e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.455730] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 694.455730] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52307af8-5579-2b2a-7ce4-84af35dfad6e" [ 694.455730] env[61964]: _type = "Task" [ 694.455730] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.463441] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52307af8-5579-2b2a-7ce4-84af35dfad6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.570109] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.570284] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.570438] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.734636] env[61964]: DEBUG oslo_concurrency.lockutils [req-a48fb610-b7da-4d8e-8175-e4945639c46c req-2b433e62-2198-4739-8596-7ee3dd095e08 service nova] Releasing lock "refresh_cache-6ced6800-db29-4766-8d83-b63b50d5fcc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.776094] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.181s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.969835] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52307af8-5579-2b2a-7ce4-84af35dfad6e, 'name': SearchDatastore_Task, 'duration_secs': 0.011367} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.973218] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-397a373e-ef31-47bd-8a42-ea9f4813f2ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.978959] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 694.978959] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c54a5-a51b-ec51-d0eb-08a54777c32e" [ 694.978959] env[61964]: _type = "Task" [ 694.978959] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.989884] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c54a5-a51b-ec51-d0eb-08a54777c32e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.010774] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b5cc9c-b4ba-45bf-b249-f45873d83ada {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.019466] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fac9a7-3500-4002-b735-36b37ff95e73 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.055540] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9908acfa-b9c0-46af-8ac7-7a6554c227e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.063726] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c0b720-2a0b-4f2a-854b-080e5f94f678 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.085754] env[61964]: DEBUG nova.compute.provider_tree [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.187026] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.278935] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 695.461192] env[61964]: DEBUG nova.network.neutron [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updating instance_info_cache with network_info: [{"id": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "address": "fa:16:3e:32:c5:75", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f7cab9-da", "ovs_interfaceid": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.490244] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c54a5-a51b-ec51-d0eb-08a54777c32e, 'name': SearchDatastore_Task, 'duration_secs': 0.019085} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.490992] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.490992] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6ced6800-db29-4766-8d83-b63b50d5fcc5/6ced6800-db29-4766-8d83-b63b50d5fcc5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 695.491171] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3451aa2f-f779-4f8a-984c-4a43e9585d2d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.497957] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 695.497957] env[61964]: value = "task-1040651" [ 695.497957] env[61964]: _type = "Task" [ 695.497957] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.506601] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040651, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.595049] env[61964]: DEBUG nova.scheduler.client.report [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 695.690250] env[61964]: DEBUG nova.compute.manager [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Received event network-vif-plugged-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 695.690250] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Acquiring lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.690391] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.690738] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.690937] env[61964]: DEBUG nova.compute.manager [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] No waiting events found dispatching network-vif-plugged-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 695.691128] env[61964]: WARNING nova.compute.manager [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Received unexpected event network-vif-plugged-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b for instance with vm_state building and task_state spawning. [ 695.691450] env[61964]: DEBUG nova.compute.manager [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Received event network-changed-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 695.691539] env[61964]: DEBUG nova.compute.manager [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Refreshing instance network info cache due to event network-changed-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 695.691674] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Acquiring lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.807685] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.964032] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.964032] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Instance network_info: |[{"id": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "address": "fa:16:3e:32:c5:75", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f7cab9-da", "ovs_interfaceid": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 695.964332] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Acquired lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.964583] env[61964]: DEBUG nova.network.neutron [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Refreshing network info cache for port 74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.965871] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:c5:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74f7cab9-da08-4246-bebd-f8b2f0ec5a3b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.973901] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating folder: Project (6cf0c31f6ac649b48cfb3205d9456483). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.975279] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e536e8fc-a53f-43f3-b9dc-6cf58f78e5b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.986961] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created folder: Project (6cf0c31f6ac649b48cfb3205d9456483) in parent group-v230360. [ 695.987213] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating folder: Instances. Parent ref: group-v230403. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 695.987495] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a694a6a1-88da-485e-9c36-6df22a4cb1fb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.996795] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created folder: Instances in parent group-v230403. [ 695.997096] env[61964]: DEBUG oslo.service.loopingcall [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.997304] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 695.997525] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7411a92-aa57-4e27-8d21-75c1674da421 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.022536] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040651, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.024064] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.024064] env[61964]: value = "task-1040654" [ 696.024064] env[61964]: _type = "Task" [ 696.024064] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.035121] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040654, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.099375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.100481] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 696.104468] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.682s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.105211] env[61964]: INFO nova.compute.claims [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.524034] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040651, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.824451} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.528021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6ced6800-db29-4766-8d83-b63b50d5fcc5/6ced6800-db29-4766-8d83-b63b50d5fcc5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 696.528021] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 696.528021] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef1f7feb-6b93-437f-9e39-1eb179dc0023 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.538568] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040654, 'name': CreateVM_Task, 'duration_secs': 0.454404} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.539977] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 696.540363] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 696.540363] env[61964]: value = "task-1040655" [ 696.540363] env[61964]: _type = "Task" [ 696.540363] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.540980] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.541166] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.541563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 696.541880] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e72132f-bb01-46f6-977a-669d8d908f23 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.553474] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040655, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.553921] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 696.553921] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52598712-8837-e435-1301-a2f8c05e0cb5" [ 696.553921] env[61964]: _type = "Task" [ 696.553921] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.563303] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52598712-8837-e435-1301-a2f8c05e0cb5, 'name': SearchDatastore_Task, 'duration_secs': 0.010713} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.563662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.563899] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.564362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.564605] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.564818] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.565113] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7255cb15-86cc-4d77-8741-8836a657583c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.578827] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.578827] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 696.579487] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dfb8769-6f6e-476f-aea4-d18c3f11f131 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.585683] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 696.585683] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c99b7e-fc67-7d7d-0c47-3f59f950861c" [ 696.585683] env[61964]: _type = "Task" [ 696.585683] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.596709] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c99b7e-fc67-7d7d-0c47-3f59f950861c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.612440] env[61964]: DEBUG nova.compute.utils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.613772] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 696.613981] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 696.680161] env[61964]: DEBUG nova.policy [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f964ed45226a4c3690321775e8d3c100', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '787322c793384f849d0a2acae27a2052', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.813070] env[61964]: DEBUG nova.network.neutron [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updated VIF entry in instance network info cache for port 74f7cab9-da08-4246-bebd-f8b2f0ec5a3b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 696.813491] env[61964]: DEBUG nova.network.neutron [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updating instance_info_cache with network_info: [{"id": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "address": "fa:16:3e:32:c5:75", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f7cab9-da", "ovs_interfaceid": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.028233] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Successfully created port: 41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.053819] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040655, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066836} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.054199] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 697.055058] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3901b850-5708-4be7-9a5c-86d91753ef29 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.083513] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 6ced6800-db29-4766-8d83-b63b50d5fcc5/6ced6800-db29-4766-8d83-b63b50d5fcc5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 697.084251] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76939ed5-0d6e-417a-94be-e1bde60d8b56 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.110289] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c99b7e-fc67-7d7d-0c47-3f59f950861c, 'name': SearchDatastore_Task, 'duration_secs': 0.011916} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.112145] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 697.112145] env[61964]: value = "task-1040656" [ 697.112145] env[61964]: _type = "Task" [ 697.112145] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.112349] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b5dc489-30b3-4e33-a44a-89c844a4063d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.122329] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 697.128041] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 697.128041] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257c691-855c-c06e-91fd-5d36e199c018" [ 697.128041] env[61964]: _type = "Task" [ 697.128041] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.132420] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040656, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.139650] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257c691-855c-c06e-91fd-5d36e199c018, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.317349] env[61964]: DEBUG oslo_concurrency.lockutils [req-0d6f43f1-8529-467e-bdfe-8184fc48dfcd req-d92b1db5-a1a1-47d6-9f9d-f358f180fda0 service nova] Releasing lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.501023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c048ad95-1a46-4956-b551-e6bea7578df1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.507709] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51718691-b781-4101-b50d-fec9ba5870db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.538015] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e6f9ca-26b6-4e7c-84c2-97db12d33dd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.545482] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3f1d59-81c8-44ef-87ca-c6d84cb3fed0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.559734] env[61964]: DEBUG nova.compute.provider_tree [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.626232] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.643197] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257c691-855c-c06e-91fd-5d36e199c018, 'name': SearchDatastore_Task, 'duration_secs': 0.022508} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.643413] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.643758] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] aef4c3a7-641a-4356-9187-ae4c082ccde9/aef4c3a7-641a-4356-9187-ae4c082ccde9.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 697.644132] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc5bd208-7962-40bc-86b5-77af4ece94f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.651138] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 697.651138] env[61964]: value = "task-1040657" [ 697.651138] env[61964]: _type = "Task" [ 697.651138] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.660599] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040657, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.063969] env[61964]: DEBUG nova.scheduler.client.report [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 698.129859] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040656, 'name': ReconfigVM_Task, 'duration_secs': 0.96406} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.130022] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 6ced6800-db29-4766-8d83-b63b50d5fcc5/6ced6800-db29-4766-8d83-b63b50d5fcc5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.131166] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdee1219-28e4-4857-a289-1822c00cfbec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.137985] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 698.142154] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 698.142154] env[61964]: value = "task-1040658" [ 698.142154] env[61964]: _type = "Task" [ 698.142154] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.160140] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040658, 'name': Rename_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.166766] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040657, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.178101] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 698.178457] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.178811] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.179082] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.179307] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.179522] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 698.179778] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 698.179958] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 698.180153] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 698.180369] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 698.180604] env[61964]: DEBUG nova.virt.hardware [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 698.181673] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55528b70-3255-4291-9714-fee2daaf082c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.191383] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc8f356-0473-4306-b8f8-70c37977066e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.477428] env[61964]: DEBUG nova.compute.manager [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Received event network-vif-plugged-41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 698.477651] env[61964]: DEBUG oslo_concurrency.lockutils [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] Acquiring lock "16276be9-b305-4d1d-afde-bc98be42687a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.477868] env[61964]: DEBUG oslo_concurrency.lockutils [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] Lock "16276be9-b305-4d1d-afde-bc98be42687a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.478108] env[61964]: DEBUG oslo_concurrency.lockutils [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] Lock "16276be9-b305-4d1d-afde-bc98be42687a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.478331] env[61964]: DEBUG nova.compute.manager [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] No waiting events found dispatching network-vif-plugged-41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 698.478566] env[61964]: WARNING nova.compute.manager [req-53e3e7ff-24bc-4ae3-ac4e-5e65e954b645 req-53a3bee9-cb43-44fc-8bf0-7341792f7a4f service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Received unexpected event network-vif-plugged-41feaa65-ae5c-49a5-828a-4e85d7aecb93 for instance with vm_state building and task_state spawning. [ 698.569417] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.573018] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 698.573276] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.175s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.575039] env[61964]: INFO nova.compute.claims [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.612945] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Successfully updated port: 41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 698.653160] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040658, 'name': Rename_Task, 'duration_secs': 0.207592} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.658241] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 698.658535] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0db35f93-b929-4afa-a91c-7ad506f42de4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.665452] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040657, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709708} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.666712] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] aef4c3a7-641a-4356-9187-ae4c082ccde9/aef4c3a7-641a-4356-9187-ae4c082ccde9.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 698.667146] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.667279] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 698.667279] env[61964]: value = "task-1040659" [ 698.667279] env[61964]: _type = "Task" [ 698.667279] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.667463] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18265660-f25d-40a5-9eed-0d72b844f25d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.677779] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.679063] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 698.679063] env[61964]: value = "task-1040660" [ 698.679063] env[61964]: _type = "Task" [ 698.679063] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.686778] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.080046] env[61964]: DEBUG nova.compute.utils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.081506] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 699.081618] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.115868] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.116071] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.116263] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.120806] env[61964]: DEBUG nova.policy [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a43a916406ce41edaf40ca81e8262f9b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '785a057715604ef08d6eca97a7488bc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.180282] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040659, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.188707] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06614} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.188999] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 699.189799] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0785634c-aac4-4f82-a3b7-640c9cfde481 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.212337] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] aef4c3a7-641a-4356-9187-ae4c082ccde9/aef4c3a7-641a-4356-9187-ae4c082ccde9.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 699.212645] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33c8eb16-e369-4b15-9dc0-2fa6de18bc80 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.234287] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 699.234287] env[61964]: value = "task-1040661" [ 699.234287] env[61964]: _type = "Task" [ 699.234287] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.242717] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.441108] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Successfully created port: 6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.584722] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 699.663419] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.680761] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040659, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.745640] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.821934] env[61964]: DEBUG nova.network.neutron [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Updating instance_info_cache with network_info: [{"id": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "address": "fa:16:3e:8a:d1:58", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41feaa65-ae", "ovs_interfaceid": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.066316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077a6d30-9949-4996-a8fb-252aed77c19b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.073967] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c52be4a-acac-4672-8cf3-735263ba5a47 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.112750] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc8ba07-9792-42ad-9ee2-2d61b3574720 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.121036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34de50f-d9cf-4fe3-8f82-e6da8916951b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.134742] env[61964]: DEBUG nova.compute.provider_tree [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.181934] env[61964]: DEBUG oslo_vmware.api [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040659, 'name': PowerOnVM_Task, 'duration_secs': 1.049022} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.182241] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 700.182451] env[61964]: INFO nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Took 9.16 seconds to spawn the instance on the hypervisor. [ 700.182644] env[61964]: DEBUG nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 700.183431] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff35c15-0e39-45cf-b7b4-ddea7069b3d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.246458] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040661, 'name': ReconfigVM_Task, 'duration_secs': 0.619708} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.246781] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Reconfigured VM instance instance-00000026 to attach disk [datastore1] aef4c3a7-641a-4356-9187-ae4c082ccde9/aef4c3a7-641a-4356-9187-ae4c082ccde9.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.247451] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c110f6a7-379c-4b75-926e-f7a675178c27 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.254048] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 700.254048] env[61964]: value = "task-1040662" [ 700.254048] env[61964]: _type = "Task" [ 700.254048] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.262234] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040662, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.328460] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.328676] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Instance network_info: |[{"id": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "address": "fa:16:3e:8a:d1:58", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41feaa65-ae", "ovs_interfaceid": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 700.329159] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:d1:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '412cde91-d0f0-4193-b36b-d8b9d17384c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41feaa65-ae5c-49a5-828a-4e85d7aecb93', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.336623] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating folder: Project (787322c793384f849d0a2acae27a2052). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.336908] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8fd83a55-a6bf-4b4b-8e0c-bc4c13930112 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.347950] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created folder: Project (787322c793384f849d0a2acae27a2052) in parent group-v230360. [ 700.348189] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating folder: Instances. Parent ref: group-v230406. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.348430] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e4bb24d-6fee-4926-ba03-43df909b3e24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.358459] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created folder: Instances in parent group-v230406. [ 700.358697] env[61964]: DEBUG oslo.service.loopingcall [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.358888] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 700.359127] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-600b8890-0e78-46fe-9ca9-0997b34374e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.378140] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.378140] env[61964]: value = "task-1040665" [ 700.378140] env[61964]: _type = "Task" [ 700.378140] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.386553] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040665, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.537654] env[61964]: DEBUG nova.compute.manager [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Received event network-changed-41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 700.537852] env[61964]: DEBUG nova.compute.manager [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Refreshing instance network info cache due to event network-changed-41feaa65-ae5c-49a5-828a-4e85d7aecb93. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 700.538092] env[61964]: DEBUG oslo_concurrency.lockutils [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] Acquiring lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.538242] env[61964]: DEBUG oslo_concurrency.lockutils [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] Acquired lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.538403] env[61964]: DEBUG nova.network.neutron [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Refreshing network info cache for port 41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.613429] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 700.639559] env[61964]: DEBUG nova.scheduler.client.report [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.650267] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 700.650267] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.650480] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 700.650747] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.650965] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 700.651156] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 700.651369] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 700.651578] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 700.651777] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 700.652064] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 700.652325] env[61964]: DEBUG nova.virt.hardware [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 700.653240] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ba8372-f4ff-4b5f-9dbd-0db3a7448a8e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.662301] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47f8a13-e851-4aaf-8c14-586fbc5ed732 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.700895] env[61964]: INFO nova.compute.manager [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Took 43.45 seconds to build instance. [ 700.764352] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040662, 'name': Rename_Task, 'duration_secs': 0.172016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.764573] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 700.764853] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0465a1a-06a4-4392-a0c3-e62739c094ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.770359] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 700.770359] env[61964]: value = "task-1040666" [ 700.770359] env[61964]: _type = "Task" [ 700.770359] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.777615] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040666, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.888955] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040665, 'name': CreateVM_Task, 'duration_secs': 0.363836} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.889151] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 700.889848] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.890041] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.890378] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 700.890639] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcb51c0f-64d5-40d6-8305-0798a90ca884 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.895305] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 700.895305] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5270f61b-9c73-66a5-6bd3-a13d9c68ac9d" [ 700.895305] env[61964]: _type = "Task" [ 700.895305] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.903086] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5270f61b-9c73-66a5-6bd3-a13d9c68ac9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.144255] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.144804] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 701.147816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.963s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.148160] env[61964]: DEBUG nova.objects.instance [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'resources' on Instance uuid 0f1ed823-1531-4505-a68a-cb172f804655 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 701.203823] env[61964]: DEBUG oslo_concurrency.lockutils [None req-20a3134f-41c8-4c6a-83f7-1604b45321a7 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.568s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.279462] env[61964]: DEBUG oslo_vmware.api [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040666, 'name': PowerOnVM_Task, 'duration_secs': 0.446163} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.279798] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 701.279934] env[61964]: INFO nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Took 7.69 seconds to spawn the instance on the hypervisor. [ 701.280231] env[61964]: DEBUG nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 701.281223] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef661879-e1e9-4a9e-a3d0-47d1debc9385 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.321741] env[61964]: DEBUG nova.network.neutron [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Updated VIF entry in instance network info cache for port 41feaa65-ae5c-49a5-828a-4e85d7aecb93. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 701.322136] env[61964]: DEBUG nova.network.neutron [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Updating instance_info_cache with network_info: [{"id": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "address": "fa:16:3e:8a:d1:58", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41feaa65-ae", "ovs_interfaceid": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.406058] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5270f61b-9c73-66a5-6bd3-a13d9c68ac9d, 'name': SearchDatastore_Task, 'duration_secs': 0.010605} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.406374] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.406611] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.407029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.407199] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.407384] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.407637] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d211aba0-a9c4-4bb9-9701-6897536235a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.415832] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.416062] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 701.416703] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9bda077-67b4-4ee8-97b5-802dcc0cf6b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.422667] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 701.422667] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520863c9-af79-8cd5-0d15-ededca67f26b" [ 701.422667] env[61964]: _type = "Task" [ 701.422667] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.429926] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520863c9-af79-8cd5-0d15-ededca67f26b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.505815] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Successfully updated port: 6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.652167] env[61964]: DEBUG nova.compute.utils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.652580] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 701.652760] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.706152] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 701.719420] env[61964]: DEBUG nova.policy [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be6009f80ede4015908df2174a246cfb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a777c3c9df5a450381e5ecca517694eb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.801718] env[61964]: INFO nova.compute.manager [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Took 38.87 seconds to build instance. [ 701.829487] env[61964]: DEBUG oslo_concurrency.lockutils [req-5030c94d-55f3-4cac-a19b-f649414e9004 req-3b75e3a5-8a2d-4635-b1f1-32829f9f9f8b service nova] Releasing lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.933817] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520863c9-af79-8cd5-0d15-ededca67f26b, 'name': SearchDatastore_Task, 'duration_secs': 0.013376} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.934861] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9039728-8648-410e-9638-06c2291d5914 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.947195] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 701.947195] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c2a944-ac10-6a3d-e33f-3e612abb812c" [ 701.947195] env[61964]: _type = "Task" [ 701.947195] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.960769] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c2a944-ac10-6a3d-e33f-3e612abb812c, 'name': SearchDatastore_Task, 'duration_secs': 0.014427} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.960769] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.960920] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/16276be9-b305-4d1d-afde-bc98be42687a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.961180] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1cdc498-c505-49a1-a357-dbc6c5651aa7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.971296] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 701.971296] env[61964]: value = "task-1040667" [ 701.971296] env[61964]: _type = "Task" [ 701.971296] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.982915] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.007964] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.008212] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquired lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.008395] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.124770] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ae7506-cabb-4937-8b6d-a7636b67394a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.133307] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a09819d-b076-462b-8c4a-1c18982c7829 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.167581] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 702.171781] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Successfully created port: d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.174520] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c132be8-1755-4543-9a17-68f3cc753605 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.186863] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ef9570-aa19-4519-a205-a6814561da15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.207705] env[61964]: DEBUG nova.compute.provider_tree [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.228346] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.307239] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e745185b-69e6-464c-834f-13e3baf7a23c tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.158s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.470914] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.471252] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.471507] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.471734] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.471924] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.478162] env[61964]: INFO nova.compute.manager [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Terminating instance [ 702.486047] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040667, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478926} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.486377] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/16276be9-b305-4d1d-afde-bc98be42687a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 702.486880] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.488173] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69a63052-da20-4d28-8e0a-2957c9adf1ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.498020] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 702.498020] env[61964]: value = "task-1040668" [ 702.498020] env[61964]: _type = "Task" [ 702.498020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.506537] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040668, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.552218] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.552534] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.552773] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.552971] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.553193] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.555453] env[61964]: INFO nova.compute.manager [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Terminating instance [ 702.581395] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.711990] env[61964]: DEBUG nova.scheduler.client.report [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 702.770149] env[61964]: DEBUG nova.compute.manager [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Received event network-vif-plugged-6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 702.770376] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Acquiring lock "a2a25787-1426-4c09-892d-395e1becccc9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.770592] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Lock "a2a25787-1426-4c09-892d-395e1becccc9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.770769] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Lock "a2a25787-1426-4c09-892d-395e1becccc9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.770941] env[61964]: DEBUG nova.compute.manager [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] No waiting events found dispatching network-vif-plugged-6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 702.771155] env[61964]: WARNING nova.compute.manager [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Received unexpected event network-vif-plugged-6b0cb07c-91a0-4e3f-a285-82c6ee167d97 for instance with vm_state building and task_state spawning. [ 702.771330] env[61964]: DEBUG nova.compute.manager [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Received event network-changed-6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 702.771544] env[61964]: DEBUG nova.compute.manager [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Refreshing instance network info cache due to event network-changed-6b0cb07c-91a0-4e3f-a285-82c6ee167d97. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 702.771656] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Acquiring lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.800686] env[61964]: DEBUG nova.network.neutron [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Updating instance_info_cache with network_info: [{"id": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "address": "fa:16:3e:c5:d4:8b", "network": {"id": "c40acbc8-6597-4780-b5ff-b6f840e3a6ef", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-500126080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "785a057715604ef08d6eca97a7488bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b0cb07c-91", "ovs_interfaceid": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.810120] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 702.982767] env[61964]: DEBUG nova.compute.manager [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 702.983045] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 702.984067] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0252f10-fbfa-45c4-9e02-b5de7ac00346 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.992845] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 702.993124] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ade9ca23-2aae-4efb-8aed-3f22d2cd4413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.003153] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 703.003153] env[61964]: value = "task-1040669" [ 703.003153] env[61964]: _type = "Task" [ 703.003153] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.010310] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040668, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084505} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.011071] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.012116] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f733965-a8d9-4bd9-9129-a05dbcbf3150 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.018467] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.039877] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/16276be9-b305-4d1d-afde-bc98be42687a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.040243] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c100ced-240f-45ef-a8f1-6cfbb63559a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.063800] env[61964]: DEBUG nova.compute.manager [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 703.063800] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.063800] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 703.063800] env[61964]: value = "task-1040670" [ 703.063800] env[61964]: _type = "Task" [ 703.063800] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.064379] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa0607d-246a-4aa7-83d8-cfd03b024230 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.078028] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040670, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.078333] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 703.078577] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69cc0343-c886-49a8-b73c-2bbd46028dae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.086054] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 703.086054] env[61964]: value = "task-1040671" [ 703.086054] env[61964]: _type = "Task" [ 703.086054] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.095951] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.187519] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 703.218179] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.070s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.222766] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 703.222766] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.222766] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 703.223245] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.223245] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 703.223351] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 703.223534] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 703.223727] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 703.223897] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 703.224191] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 703.224284] env[61964]: DEBUG nova.virt.hardware [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 703.224633] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.282s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.224893] env[61964]: DEBUG nova.objects.instance [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lazy-loading 'resources' on Instance uuid 06631c59-9969-4763-bfe8-472e8dcf8848 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 703.226679] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4377817a-05d2-4e1d-961e-c03b0a268315 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.237146] env[61964]: INFO nova.scheduler.client.report [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted allocations for instance 0f1ed823-1531-4505-a68a-cb172f804655 [ 703.239633] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3385d982-506a-4546-b721-59ef005f2063 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.303688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Releasing lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.303888] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Instance network_info: |[{"id": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "address": "fa:16:3e:c5:d4:8b", "network": {"id": "c40acbc8-6597-4780-b5ff-b6f840e3a6ef", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-500126080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "785a057715604ef08d6eca97a7488bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b0cb07c-91", "ovs_interfaceid": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 703.304239] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Acquired lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.304439] env[61964]: DEBUG nova.network.neutron [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Refreshing network info cache for port 6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 703.305823] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:d4:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69cfa7ba-6989-4d75-9495-97b5fea00c3c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b0cb07c-91a0-4e3f-a285-82c6ee167d97', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.316022] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Creating folder: Project (785a057715604ef08d6eca97a7488bc8). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 703.320919] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa930e9c-ae0e-4126-b0a7-3dd1b2961ae1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.338260] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.338625] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Created folder: Project (785a057715604ef08d6eca97a7488bc8) in parent group-v230360. [ 703.338810] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Creating folder: Instances. Parent ref: group-v230409. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 703.339066] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff372fc4-4acd-4e64-af5d-a804410cce3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.350657] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Created folder: Instances in parent group-v230409. [ 703.350924] env[61964]: DEBUG oslo.service.loopingcall [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.351144] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 703.351354] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29c9812c-e0e6-4c54-aea1-9aee383b36ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.375856] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.375856] env[61964]: value = "task-1040674" [ 703.375856] env[61964]: _type = "Task" [ 703.375856] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.384456] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040674, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.515162] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040669, 'name': PowerOffVM_Task, 'duration_secs': 0.316881} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.515666] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 703.516025] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 703.516406] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eef76623-01eb-41ff-9b82-222f7d50be34 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.579881] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040670, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.601151] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 703.601151] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 703.601151] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleting the datastore file [datastore1] d99859dc-2206-42d9-ae6d-8294fbd6942c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.603367] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9b8fe94-4b0b-4270-8dd1-f4c84d93edba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.605503] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040671, 'name': PowerOffVM_Task, 'duration_secs': 0.266439} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.605983] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 703.607058] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 703.609771] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43805c34-45b6-4760-9fdb-9a9d6854d2c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.615091] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 703.615091] env[61964]: value = "task-1040676" [ 703.615091] env[61964]: _type = "Task" [ 703.615091] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.617934] env[61964]: DEBUG nova.network.neutron [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Updated VIF entry in instance network info cache for port 6b0cb07c-91a0-4e3f-a285-82c6ee167d97. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 703.619095] env[61964]: DEBUG nova.network.neutron [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Updating instance_info_cache with network_info: [{"id": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "address": "fa:16:3e:c5:d4:8b", "network": {"id": "c40acbc8-6597-4780-b5ff-b6f840e3a6ef", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-500126080-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "785a057715604ef08d6eca97a7488bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69cfa7ba-6989-4d75-9495-97b5fea00c3c", "external-id": "nsx-vlan-transportzone-225", "segmentation_id": 225, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b0cb07c-91", "ovs_interfaceid": "6b0cb07c-91a0-4e3f-a285-82c6ee167d97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.622832] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.758200] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ee6c923-08a6-4e2f-a87f-230ccef6388b tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "0f1ed823-1531-4505-a68a-cb172f804655" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.621s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.760810] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 703.761194] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 703.766370] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleting the datastore file [datastore1] 6ced6800-db29-4766-8d83-b63b50d5fcc5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.766370] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06f8bd96-1aef-4584-ab5e-d95889de5903 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.771859] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 703.771859] env[61964]: value = "task-1040678" [ 703.771859] env[61964]: _type = "Task" [ 703.771859] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.787471] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.886845] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040674, 'name': CreateVM_Task, 'duration_secs': 0.487397} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.887054] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 703.887763] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.887998] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.888437] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 703.888709] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a5e19da-7d1d-4a56-bf61-5821be32a659 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.893809] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 703.893809] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fe32e9-533d-f1c6-ba4a-aa9c320a0a9f" [ 703.893809] env[61964]: _type = "Task" [ 703.893809] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.908318] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fe32e9-533d-f1c6-ba4a-aa9c320a0a9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.012138] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Successfully updated port: d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.082116] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040670, 'name': ReconfigVM_Task, 'duration_secs': 0.660821} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.082372] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/16276be9-b305-4d1d-afde-bc98be42687a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.083383] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4bb067e-fd54-45d3-b8f1-9995649bdaf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.095689] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 704.095689] env[61964]: value = "task-1040679" [ 704.095689] env[61964]: _type = "Task" [ 704.095689] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.105774] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040679, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.124505] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa8b8dc7-3e0f-40db-b1a7-5e22b1e5edc4 req-60bb7ef9-ad02-4ac7-8355-88286dfd4caa service nova] Releasing lock "refresh_cache-a2a25787-1426-4c09-892d-395e1becccc9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.125017] env[61964]: DEBUG oslo_vmware.api [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.351596} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.127903] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 704.128294] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 704.128540] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 704.128764] env[61964]: INFO nova.compute.manager [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 704.129074] env[61964]: DEBUG oslo.service.loopingcall [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.129580] env[61964]: DEBUG nova.compute.manager [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.129912] env[61964]: DEBUG nova.network.neutron [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.228646] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5c9cbb-ae0b-4e1d-9b7d-fad0d40ac3a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.238012] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62f2612-4334-43df-a707-7dba0f260a3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.281030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d7e8fb-2225-44d9-bb27-11184042f814 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.286816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.287144] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.294523] env[61964]: DEBUG oslo_vmware.api [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172801} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.296682] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 704.296939] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 704.297167] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 704.297387] env[61964]: INFO nova.compute.manager [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Took 1.23 seconds to destroy the instance on the hypervisor. [ 704.297688] env[61964]: DEBUG oslo.service.loopingcall [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.297977] env[61964]: DEBUG nova.compute.manager [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.298118] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.300925] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88b029a-3aef-47b6-9755-ede57d3d474f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.318159] env[61964]: DEBUG nova.compute.provider_tree [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.407147] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fe32e9-533d-f1c6-ba4a-aa9c320a0a9f, 'name': SearchDatastore_Task, 'duration_secs': 0.01061} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.407531] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.407780] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.408061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.408202] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.408571] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.408862] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51fb9551-13a6-4ad6-9a38-7f6e417f3c8e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.419680] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.420556] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 704.421352] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be78b215-5b07-492e-ac60-0970c9eced08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.440191] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 704.440191] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cfb9f-c4f2-1fd7-5c14-c5c49af1ae8b" [ 704.440191] env[61964]: _type = "Task" [ 704.440191] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.455149] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cfb9f-c4f2-1fd7-5c14-c5c49af1ae8b, 'name': SearchDatastore_Task, 'duration_secs': 0.022016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.456476] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b527d025-5df6-4969-b18b-224d838659d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.462785] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 704.462785] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526226b6-52e1-8b9d-aa99-e34b6ede9ffe" [ 704.462785] env[61964]: _type = "Task" [ 704.462785] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.473735] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526226b6-52e1-8b9d-aa99-e34b6ede9ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.517108] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.517361] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquired lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.517554] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.585282] env[61964]: DEBUG nova.compute.manager [req-f0f7c1ea-30eb-482c-921c-472a9cd91076 req-68739804-93ab-4131-bffe-a4ba481f726d service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Received event network-vif-deleted-bd317133-fb07-49c5-80d0-008894442dc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 704.586036] env[61964]: INFO nova.compute.manager [req-f0f7c1ea-30eb-482c-921c-472a9cd91076 req-68739804-93ab-4131-bffe-a4ba481f726d service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Neutron deleted interface bd317133-fb07-49c5-80d0-008894442dc7; detaching it from the instance and deleting it from the info cache [ 704.586036] env[61964]: DEBUG nova.network.neutron [req-f0f7c1ea-30eb-482c-921c-472a9cd91076 req-68739804-93ab-4131-bffe-a4ba481f726d service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.606453] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040679, 'name': Rename_Task, 'duration_secs': 0.417614} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.606746] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 704.606996] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07a015d5-acc3-4707-8d0d-784d2f5cb8ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.614816] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 704.614816] env[61964]: value = "task-1040680" [ 704.614816] env[61964]: _type = "Task" [ 704.614816] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.623501] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.821609] env[61964]: DEBUG nova.scheduler.client.report [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 704.839711] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Received event network-vif-plugged-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 704.839944] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Acquiring lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.840181] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.840397] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.840588] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] No waiting events found dispatching network-vif-plugged-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 704.840761] env[61964]: WARNING nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Received unexpected event network-vif-plugged-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 for instance with vm_state building and task_state spawning. [ 704.840926] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Received event network-changed-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 704.841095] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Refreshing instance network info cache due to event network-changed-d1e722bd-f0de-4c54-b282-7ea69ab92fd0. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 704.841268] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Acquiring lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.968913] env[61964]: DEBUG nova.network.neutron [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.976875] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526226b6-52e1-8b9d-aa99-e34b6ede9ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.025577} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.977157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.977552] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] a2a25787-1426-4c09-892d-395e1becccc9/a2a25787-1426-4c09-892d-395e1becccc9.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 704.977673] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e892f9a-f4c9-4919-911b-b00128b2925d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.985611] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 704.985611] env[61964]: value = "task-1040681" [ 704.985611] env[61964]: _type = "Task" [ 704.985611] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.995145] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.060663] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.063042] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.088253] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bda04389-54e0-47f8-9dcf-8b32e5920b07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.099182] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe1571b-3db9-4e5c-b942-c28a64d5b3e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.126174] env[61964]: DEBUG oslo_vmware.api [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040680, 'name': PowerOnVM_Task, 'duration_secs': 0.462711} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.134910] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 705.135170] env[61964]: INFO nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Took 7.00 seconds to spawn the instance on the hypervisor. [ 705.135361] env[61964]: DEBUG nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 705.135713] env[61964]: DEBUG nova.compute.manager [req-f0f7c1ea-30eb-482c-921c-472a9cd91076 req-68739804-93ab-4131-bffe-a4ba481f726d service nova] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Detach interface failed, port_id=bd317133-fb07-49c5-80d0-008894442dc7, reason: Instance 6ced6800-db29-4766-8d83-b63b50d5fcc5 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 705.136625] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6562b980-a08c-4c12-a2e1-dde03814028a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.251383] env[61964]: DEBUG nova.network.neutron [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updating instance_info_cache with network_info: [{"id": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "address": "fa:16:3e:8e:68:89", "network": {"id": "5c28edbb-5331-40e7-889f-5512f505e058", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-593166019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a777c3c9df5a450381e5ecca517694eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e722bd-f0", "ovs_interfaceid": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.327568] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.330910] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.149s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.333384] env[61964]: INFO nova.compute.claims [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.366870] env[61964]: INFO nova.scheduler.client.report [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Deleted allocations for instance 06631c59-9969-4763-bfe8-472e8dcf8848 [ 705.474078] env[61964]: INFO nova.compute.manager [-] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Took 1.34 seconds to deallocate network for instance. [ 705.504030] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040681, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474626} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.504329] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] a2a25787-1426-4c09-892d-395e1becccc9/a2a25787-1426-4c09-892d-395e1becccc9.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 705.504557] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 705.504852] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-849d9603-f741-4bba-82e6-6276d9adb38c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.515811] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 705.515811] env[61964]: value = "task-1040682" [ 705.515811] env[61964]: _type = "Task" [ 705.515811] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.526803] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040682, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.565671] env[61964]: INFO nova.compute.manager [-] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Took 1.27 seconds to deallocate network for instance. [ 705.653741] env[61964]: INFO nova.compute.manager [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Took 33.19 seconds to build instance. [ 705.754152] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Releasing lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.754539] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Instance network_info: |[{"id": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "address": "fa:16:3e:8e:68:89", "network": {"id": "5c28edbb-5331-40e7-889f-5512f505e058", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-593166019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a777c3c9df5a450381e5ecca517694eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e722bd-f0", "ovs_interfaceid": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 705.754936] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Acquired lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.755182] env[61964]: DEBUG nova.network.neutron [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Refreshing network info cache for port d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.756325] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:68:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1e722bd-f0de-4c54-b282-7ea69ab92fd0', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 705.765032] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Creating folder: Project (a777c3c9df5a450381e5ecca517694eb). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.765608] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57b9d48f-3b5c-46ba-ad25-6e7f8bbae1ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.779381] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Created folder: Project (a777c3c9df5a450381e5ecca517694eb) in parent group-v230360. [ 705.779579] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Creating folder: Instances. Parent ref: group-v230412. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.779819] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71dc06a4-75b5-490e-b9e1-e3a31ef4f8d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.788527] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Created folder: Instances in parent group-v230412. [ 705.788758] env[61964]: DEBUG oslo.service.loopingcall [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.788948] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 705.789169] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cf5a9a8-0118-4060-bf3d-13c9c7807fec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.808708] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 705.808708] env[61964]: value = "task-1040685" [ 705.808708] env[61964]: _type = "Task" [ 705.808708] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.817227] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040685, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.881435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2218cd3e-8579-4705-907d-538ca43dd3f0 tempest-ServerDiagnosticsV248Test-1665881153 tempest-ServerDiagnosticsV248Test-1665881153-project-member] Lock "06631c59-9969-4763-bfe8-472e8dcf8848" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.335s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.980210] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.027961] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040682, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066967} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.028278] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.029125] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee0ccd8-ac6b-4203-b5cd-0fdd56f564a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.057370] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] a2a25787-1426-4c09-892d-395e1becccc9/a2a25787-1426-4c09-892d-395e1becccc9.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.057370] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8632f630-34b0-4a60-8022-11005864122d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.072129] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.082028] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 706.082028] env[61964]: value = "task-1040686" [ 706.082028] env[61964]: _type = "Task" [ 706.082028] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.091426] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.156056] env[61964]: DEBUG oslo_concurrency.lockutils [None req-823b3ecc-6bcb-4d3a-979c-3a7fe27f1210 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.345s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.319335] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040685, 'name': CreateVM_Task, 'duration_secs': 0.376769} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.319694] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 706.320203] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.320368] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.320686] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 706.320942] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8814919-c444-4ebd-bb43-af418f5f0a43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.325952] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 706.325952] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cee89c-9f92-008d-53dc-9547fea45c89" [ 706.325952] env[61964]: _type = "Task" [ 706.325952] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.334559] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cee89c-9f92-008d-53dc-9547fea45c89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.542766] env[61964]: DEBUG nova.network.neutron [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updated VIF entry in instance network info cache for port d1e722bd-f0de-4c54-b282-7ea69ab92fd0. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 706.543220] env[61964]: DEBUG nova.network.neutron [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updating instance_info_cache with network_info: [{"id": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "address": "fa:16:3e:8e:68:89", "network": {"id": "5c28edbb-5331-40e7-889f-5512f505e058", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-593166019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a777c3c9df5a450381e5ecca517694eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e722bd-f0", "ovs_interfaceid": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.595564] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.658574] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 706.697029] env[61964]: INFO nova.compute.manager [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Rescuing [ 706.697029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.697029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.697029] env[61964]: DEBUG nova.network.neutron [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.782443] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c44571c-191f-429a-bf31-34be8e8f6cd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.790695] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d6a9d2-c6a4-4dae-85d8-79e1f6b896cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.823076] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2864e4cf-85c1-4379-b3ec-012f318558a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.833455] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2af05b-96d5-43b6-b58f-39d39a88dd2d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.841308] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cee89c-9f92-008d-53dc-9547fea45c89, 'name': SearchDatastore_Task, 'duration_secs': 0.01026} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.841994] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.842273] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 706.843075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.843075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.843075] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 706.843236] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25fdb40b-4b85-4a0c-9a1f-3f25daab205d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.852810] env[61964]: DEBUG nova.compute.provider_tree [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.861390] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 706.861583] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 706.862352] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-916d7d17-4f2b-4e8a-bbfa-97b05600e38f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.868272] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 706.868272] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288c58f-d988-be47-de8c-8337d3fef516" [ 706.868272] env[61964]: _type = "Task" [ 706.868272] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.877102] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288c58f-d988-be47-de8c-8337d3fef516, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.046561] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] Releasing lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.046979] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Received event network-vif-deleted-f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 707.047322] env[61964]: INFO nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Neutron deleted interface f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433; detaching it from the instance and deleting it from the info cache [ 707.047574] env[61964]: DEBUG nova.network.neutron [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.094435] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040686, 'name': ReconfigVM_Task, 'duration_secs': 0.858203} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.094435] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Reconfigured VM instance instance-00000028 to attach disk [datastore2] a2a25787-1426-4c09-892d-395e1becccc9/a2a25787-1426-4c09-892d-395e1becccc9.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.095823] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b7dcf41-e4b3-45ee-9b1a-b1dea9819491 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.103044] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 707.103044] env[61964]: value = "task-1040687" [ 707.103044] env[61964]: _type = "Task" [ 707.103044] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.111966] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040687, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.183379] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.357244] env[61964]: DEBUG nova.scheduler.client.report [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 707.380035] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288c58f-d988-be47-de8c-8337d3fef516, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.381440] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9bbb6ca-653d-4aa7-a075-496b0a1f5374 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.388074] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 707.388074] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52671358-6208-b321-1a43-e415c1ea726b" [ 707.388074] env[61964]: _type = "Task" [ 707.388074] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.398759] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52671358-6208-b321-1a43-e415c1ea726b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.438207] env[61964]: DEBUG nova.network.neutron [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Updating instance_info_cache with network_info: [{"id": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "address": "fa:16:3e:8a:d1:58", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41feaa65-ae", "ovs_interfaceid": "41feaa65-ae5c-49a5-828a-4e85d7aecb93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.550745] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3d6a576-ce23-47f5-a456-eba6d49362b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.562022] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41775ba-6977-4c02-b766-b776435b7071 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.593281] env[61964]: DEBUG nova.compute.manager [req-1c16ac17-f429-485b-9c3f-a3ffd3f1be07 req-eeb2fe53-997c-4328-8a06-6776d3ed5f9d service nova] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Detach interface failed, port_id=f6610a7b-e3c0-4ee0-a3cc-1fdca1ce6433, reason: Instance d99859dc-2206-42d9-ae6d-8294fbd6942c could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 707.613559] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040687, 'name': Rename_Task, 'duration_secs': 0.303433} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.613912] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 707.614214] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c4349c3-3bd4-45b4-a2fe-162ec93b46e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.621654] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 707.621654] env[61964]: value = "task-1040688" [ 707.621654] env[61964]: _type = "Task" [ 707.621654] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.631384] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.865019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.865019] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 707.866767] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.546s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.868623] env[61964]: INFO nova.compute.claims [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.902026] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52671358-6208-b321-1a43-e415c1ea726b, 'name': SearchDatastore_Task, 'duration_secs': 0.009876} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.902026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.902026] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] af413b1d-9e7e-43d0-a39a-fb9dda9cf281/af413b1d-9e7e-43d0-a39a-fb9dda9cf281.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 707.902026] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-840c171f-88cc-4849-977c-770cc63b7c72 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.909881] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 707.909881] env[61964]: value = "task-1040689" [ 707.909881] env[61964]: _type = "Task" [ 707.909881] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.920342] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.942652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "refresh_cache-16276be9-b305-4d1d-afde-bc98be42687a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.134803] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040688, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.374255] env[61964]: DEBUG nova.compute.utils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.378290] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 708.378488] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 708.420965] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474415} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.421323] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] af413b1d-9e7e-43d0-a39a-fb9dda9cf281/af413b1d-9e7e-43d0-a39a-fb9dda9cf281.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 708.421500] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 708.421756] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cde92829-acdb-4dad-90f0-8febb8a7d31f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.427267] env[61964]: DEBUG nova.policy [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50f6de62a8794cbe9ef9b9bc984fa9c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98ad9397db1f4654a93d283e05664310', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.441304] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 708.441304] env[61964]: value = "task-1040690" [ 708.441304] env[61964]: _type = "Task" [ 708.441304] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.449980] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.633217] env[61964]: DEBUG oslo_vmware.api [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040688, 'name': PowerOnVM_Task, 'duration_secs': 0.561351} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.634031] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 708.634031] env[61964]: INFO nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Took 8.02 seconds to spawn the instance on the hypervisor. [ 708.634096] env[61964]: DEBUG nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 708.635019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0fb85c-2556-47f7-9fa2-ed388c20f857 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.752676] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Successfully created port: bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.879644] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 708.957831] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082121} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.960747] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.962472] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5c0522-9d09-40d1-a02c-c335b57cb8b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.991905] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] af413b1d-9e7e-43d0-a39a-fb9dda9cf281/af413b1d-9e7e-43d0-a39a-fb9dda9cf281.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.996199] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d20c474f-d294-457e-8d3e-f32e01e44ef2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.018031] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 709.018031] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2aba6ab2-840d-4b44-bfd1-8ddf2330609c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.026472] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 709.026472] env[61964]: value = "task-1040691" [ 709.026472] env[61964]: _type = "Task" [ 709.026472] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.028218] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 709.028218] env[61964]: value = "task-1040692" [ 709.028218] env[61964]: _type = "Task" [ 709.028218] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.045391] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.153021] env[61964]: INFO nova.compute.manager [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Took 34.75 seconds to build instance. [ 709.367313] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffeb9cde-2b52-49f0-823f-8c8ed0576f5d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.375643] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9981edfe-3eb0-4abc-a55c-7d41d530dc10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.409142] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a3674f-3797-432d-8ad5-ea8c69e62db9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.418450] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609e1d1b-3a9a-4a35-98c5-e9b7fe14e87b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.434458] env[61964]: DEBUG nova.compute.provider_tree [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.542386] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040691, 'name': PowerOffVM_Task, 'duration_secs': 0.340235} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.545612] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 709.545962] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040692, 'name': ReconfigVM_Task, 'duration_secs': 0.343083} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.546678] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510b77a8-ad9d-4367-a38c-42cb8fbf514f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.549106] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Reconfigured VM instance instance-00000029 to attach disk [datastore2] af413b1d-9e7e-43d0-a39a-fb9dda9cf281/af413b1d-9e7e-43d0-a39a-fb9dda9cf281.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 709.549711] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6526bf6a-9987-498e-a49f-46c0ac913099 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.569459] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d5ec4c-f0f9-4c5b-9b6b-95ea2ea080c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.572321] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 709.572321] env[61964]: value = "task-1040693" [ 709.572321] env[61964]: _type = "Task" [ 709.572321] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.589853] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040693, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.606563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 709.606950] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c38c93c-047a-4793-b29d-39f20d59c125 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.615580] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 709.615580] env[61964]: value = "task-1040694" [ 709.615580] env[61964]: _type = "Task" [ 709.615580] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.626562] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 709.626788] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.627054] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.627216] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.627402] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.627760] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ed04bee-ba9b-4b16-b98c-0accd19c6c70 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.636391] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.636590] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 709.638151] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ad59119-b466-4dc5-a868-86a937dfa9ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.644459] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 709.644459] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5241377e-34ab-6deb-f92c-e334a94aec18" [ 709.644459] env[61964]: _type = "Task" [ 709.644459] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.653456] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5241377e-34ab-6deb-f92c-e334a94aec18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.656059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9f8b1bb1-563f-4663-aed6-7034f8186aee tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.784s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.914506] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 709.937029] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 709.937359] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.937778] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 709.938221] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.938398] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 709.938561] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 709.938942] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 709.939203] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 709.939403] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 709.939578] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 709.939757] env[61964]: DEBUG nova.virt.hardware [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 709.940982] env[61964]: DEBUG nova.scheduler.client.report [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.945184] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3c7866-3523-4275-9b09-37c63b2a407e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.955045] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee697f8a-8cd3-44bc-8799-097092b711d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.082886] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040693, 'name': Rename_Task, 'duration_secs': 0.166592} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.083271] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 710.083567] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ec33b57-afc1-4212-ac9c-a38a7e8d9b9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.090530] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 710.090530] env[61964]: value = "task-1040695" [ 710.090530] env[61964]: _type = "Task" [ 710.090530] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.099568] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.155672] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5241377e-34ab-6deb-f92c-e334a94aec18, 'name': SearchDatastore_Task, 'duration_secs': 0.017456} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.156543] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5b5d8e6-cfe6-43b4-98fd-dadaa9e907cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.159245] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 710.167120] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 710.167120] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b85359-50fe-1031-bc2f-acc93da231c6" [ 710.167120] env[61964]: _type = "Task" [ 710.167120] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.176711] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b85359-50fe-1031-bc2f-acc93da231c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.379361] env[61964]: DEBUG nova.compute.manager [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Received event network-vif-plugged-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 710.379361] env[61964]: DEBUG oslo_concurrency.lockutils [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] Acquiring lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.379361] env[61964]: DEBUG oslo_concurrency.lockutils [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.379361] env[61964]: DEBUG oslo_concurrency.lockutils [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.379361] env[61964]: DEBUG nova.compute.manager [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] No waiting events found dispatching network-vif-plugged-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 710.379837] env[61964]: WARNING nova.compute.manager [req-05815142-0fc8-4df2-a478-741480ed89a3 req-9a2d11d1-4e24-49dd-9849-cbab5912527e service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Received unexpected event network-vif-plugged-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 for instance with vm_state building and task_state spawning. [ 710.422537] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Successfully updated port: bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 710.449995] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.450723] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 710.454183] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.298s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.455681] env[61964]: INFO nova.compute.claims [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.487524] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "interface-a2a25787-1426-4c09-892d-395e1becccc9-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.487695] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "interface-a2a25787-1426-4c09-892d-395e1becccc9-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.487976] env[61964]: DEBUG nova.objects.instance [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lazy-loading 'flavor' on Instance uuid a2a25787-1426-4c09-892d-395e1becccc9 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.602845] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040695, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.681684] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b85359-50fe-1031-bc2f-acc93da231c6, 'name': SearchDatastore_Task, 'duration_secs': 0.015282} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.682085] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.682364] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. {{(pid=61964) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 710.682635] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31f2000a-99bd-4b62-8d5b-694fc89901a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.685636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.691738] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 710.691738] env[61964]: value = "task-1040696" [ 710.691738] env[61964]: _type = "Task" [ 710.691738] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.702546] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.925067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.925067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquired lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.925413] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 710.960277] env[61964]: DEBUG nova.compute.utils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.966088] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 710.966088] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 710.992928] env[61964]: DEBUG nova.objects.instance [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lazy-loading 'pci_requests' on Instance uuid a2a25787-1426-4c09-892d-395e1becccc9 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.012893] env[61964]: DEBUG nova.policy [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f438e739578a4fabad898ebd5e2925cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ccf1fe5003e484fa73ce5c32767479f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.102835] env[61964]: DEBUG oslo_vmware.api [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040695, 'name': PowerOnVM_Task, 'duration_secs': 0.525269} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.103135] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 711.103430] env[61964]: INFO nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Took 7.92 seconds to spawn the instance on the hypervisor. [ 711.103524] env[61964]: DEBUG nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 711.104464] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1220038b-5b43-47a0-94bd-ba848148968f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.209029] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040696, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.317206] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Successfully created port: 6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.464727] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 711.468023] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.499156] env[61964]: DEBUG nova.objects.base [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 711.499156] env[61964]: DEBUG nova.network.neutron [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.631602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de39f828-31c5-4746-9517-263353e28a8e tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "interface-a2a25787-1426-4c09-892d-395e1becccc9-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.144s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.633078] env[61964]: INFO nova.compute.manager [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Took 35.25 seconds to build instance. [ 711.698343] env[61964]: DEBUG nova.network.neutron [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Updating instance_info_cache with network_info: [{"id": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "address": "fa:16:3e:5c:78:40", "network": {"id": "500cf95c-2c61-497b-a964-0a82f0ea8e8e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-731803685-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ad9397db1f4654a93d283e05664310", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdaf1e74-24", "ovs_interfaceid": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.706777] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612717} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.707679] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. [ 711.708621] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb7cdee-9ed1-4489-afb3-0481736e22f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.740274] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.743561] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e32c86e4-a55b-4e84-ba6b-ca3e5ba9d361 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.764131] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 711.764131] env[61964]: value = "task-1040697" [ 711.764131] env[61964]: _type = "Task" [ 711.764131] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.774596] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040697, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.920522] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13cab8e-dd8c-48b0-9eb6-6f197329e74a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.929173] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa77474-ad3e-405e-87e2-c9fdf8cb64a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.961858] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e91a5cd-e5d0-4ffa-b538-17b932c348e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.970011] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec8e483-f03a-4565-a5f1-74a26976efb5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.986858] env[61964]: DEBUG nova.compute.provider_tree [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.135880] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ac7579f3-2a2d-4d49-88fa-c4883f994049 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.410s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.208390] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Releasing lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.208785] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Instance network_info: |[{"id": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "address": "fa:16:3e:5c:78:40", "network": {"id": "500cf95c-2c61-497b-a964-0a82f0ea8e8e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-731803685-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ad9397db1f4654a93d283e05664310", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdaf1e74-24", "ovs_interfaceid": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 712.209302] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:78:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bdaf1e74-2420-4ca8-9dfe-ba0b13624e15', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.217863] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Creating folder: Project (98ad9397db1f4654a93d283e05664310). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 712.218408] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f20e1cb-c0dc-4bd1-8f20-fa25a084f94a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.233998] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Created folder: Project (98ad9397db1f4654a93d283e05664310) in parent group-v230360. [ 712.233998] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Creating folder: Instances. Parent ref: group-v230415. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 712.234214] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26d77d19-3338-48c1-a3aa-6b149f55a581 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.245020] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Created folder: Instances in parent group-v230415. [ 712.245358] env[61964]: DEBUG oslo.service.loopingcall [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.245564] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 712.245773] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fdc45b03-9190-49ee-ba73-37a9d3672d1f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.265527] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 712.265527] env[61964]: value = "task-1040700" [ 712.265527] env[61964]: _type = "Task" [ 712.265527] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.276436] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040697, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.279340] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040700, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.479135] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 712.489904] env[61964]: DEBUG nova.scheduler.client.report [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 712.512595] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 712.512595] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.512830] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 712.512867] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.513056] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 712.513234] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 712.513475] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 712.513669] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 712.513874] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 712.514464] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 712.514933] env[61964]: DEBUG nova.virt.hardware [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 712.515933] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab1e237-30a0-4aaa-bae2-82f1a4132f18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.526924] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21af009-ea81-4d82-a9b0-c3384c225359 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.640453] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 712.778921] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040697, 'name': ReconfigVM_Task, 'duration_secs': 0.765859} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.782716] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.783062] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040700, 'name': CreateVM_Task, 'duration_secs': 0.343981} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.783785] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd12312-c7b7-4889-b819-29e22ecee551 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.786243] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 712.786929] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.787113] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.787441] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 712.788058] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebfe3cea-e9f0-4770-9925-e657077f7418 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.814655] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85cdc23b-2de1-4068-8fc4-38e3ca7247e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.823817] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 712.823817] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b34a91-cd8d-fc06-3060-c023a7c90fc2" [ 712.823817] env[61964]: _type = "Task" [ 712.823817] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.830527] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 712.830527] env[61964]: value = "task-1040701" [ 712.830527] env[61964]: _type = "Task" [ 712.830527] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.837372] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b34a91-cd8d-fc06-3060-c023a7c90fc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.844252] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.907608] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Successfully updated port: 6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.995473] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.995945] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 713.003460] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 20.796s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.125204] env[61964]: DEBUG nova.compute.manager [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Received event network-changed-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 713.125204] env[61964]: DEBUG nova.compute.manager [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Refreshing instance network info cache due to event network-changed-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 713.125500] env[61964]: DEBUG oslo_concurrency.lockutils [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] Acquiring lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.125500] env[61964]: DEBUG oslo_concurrency.lockutils [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] Acquired lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.126247] env[61964]: DEBUG nova.network.neutron [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Refreshing network info cache for port bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.169721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.335670] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b34a91-cd8d-fc06-3060-c023a7c90fc2, 'name': SearchDatastore_Task, 'duration_secs': 0.072553} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.339820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.340168] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.340464] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.340657] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.340883] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.341251] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7be4a1c0-2821-4c5c-a53e-e339f1536cf3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.351909] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040701, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.354462] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.354685] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 713.355735] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-157fec82-4ad9-4a0e-9425-04f13c083402 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.362768] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 713.362768] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f5988-db3b-6be9-eb52-9bd7c8ffc684" [ 713.362768] env[61964]: _type = "Task" [ 713.362768] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.377462] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f5988-db3b-6be9-eb52-9bd7c8ffc684, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.409559] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.409724] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.410089] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.430847] env[61964]: DEBUG nova.compute.manager [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Received event network-changed-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 713.431105] env[61964]: DEBUG nova.compute.manager [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Refreshing instance network info cache due to event network-changed-d1e722bd-f0de-4c54-b282-7ea69ab92fd0. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 713.431328] env[61964]: DEBUG oslo_concurrency.lockutils [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] Acquiring lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.431470] env[61964]: DEBUG oslo_concurrency.lockutils [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] Acquired lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.431628] env[61964]: DEBUG nova.network.neutron [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Refreshing network info cache for port d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.509820] env[61964]: INFO nova.compute.claims [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.514624] env[61964]: DEBUG nova.compute.utils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.516254] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 713.516459] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.572860] env[61964]: DEBUG nova.policy [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '430778d9f17e4579a81902608fb3501a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b627b25af0c4982a22c8c0bd28641a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.846404] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040701, 'name': ReconfigVM_Task, 'duration_secs': 0.851729} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.846527] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 713.846757] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a0fbc93-5811-4203-ad22-6dbc2f78a418 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.855222] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 713.855222] env[61964]: value = "task-1040702" [ 713.855222] env[61964]: _type = "Task" [ 713.855222] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.864587] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.874251] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f5988-db3b-6be9-eb52-9bd7c8ffc684, 'name': SearchDatastore_Task, 'duration_secs': 0.012407} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.875174] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-962e0011-0e33-45d6-8715-a896dc2bc85f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.881979] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 713.881979] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e030fa-2c50-5d2a-796d-661bf7938e24" [ 713.881979] env[61964]: _type = "Task" [ 713.881979] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.891519] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e030fa-2c50-5d2a-796d-661bf7938e24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.905186] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "a2a25787-1426-4c09-892d-395e1becccc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.905628] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.905813] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "a2a25787-1426-4c09-892d-395e1becccc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.906061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.906293] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.909264] env[61964]: INFO nova.compute.manager [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Terminating instance [ 713.939846] env[61964]: DEBUG nova.network.neutron [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Updated VIF entry in instance network info cache for port bdaf1e74-2420-4ca8-9dfe-ba0b13624e15. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 713.940226] env[61964]: DEBUG nova.network.neutron [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Updating instance_info_cache with network_info: [{"id": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "address": "fa:16:3e:5c:78:40", "network": {"id": "500cf95c-2c61-497b-a964-0a82f0ea8e8e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-731803685-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ad9397db1f4654a93d283e05664310", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdaf1e74-24", "ovs_interfaceid": "bdaf1e74-2420-4ca8-9dfe-ba0b13624e15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.956060] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.023137] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 714.027475] env[61964]: INFO nova.compute.resource_tracker [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating resource usage from migration 0bbcc5dd-4fc7-4431-895f-e2476f66bace [ 714.085717] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Successfully created port: 8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.285230] env[61964]: DEBUG nova.network.neutron [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Updating instance_info_cache with network_info: [{"id": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "address": "fa:16:3e:19:59:60", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d1d176f-64", "ovs_interfaceid": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.315020] env[61964]: DEBUG nova.network.neutron [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updated VIF entry in instance network info cache for port d1e722bd-f0de-4c54-b282-7ea69ab92fd0. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 714.315595] env[61964]: DEBUG nova.network.neutron [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updating instance_info_cache with network_info: [{"id": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "address": "fa:16:3e:8e:68:89", "network": {"id": "5c28edbb-5331-40e7-889f-5512f505e058", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-593166019-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a777c3c9df5a450381e5ecca517694eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1e722bd-f0", "ovs_interfaceid": "d1e722bd-f0de-4c54-b282-7ea69ab92fd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.369590] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040702, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.396340] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e030fa-2c50-5d2a-796d-661bf7938e24, 'name': SearchDatastore_Task, 'duration_secs': 0.011579} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.399655] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.399655] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 5c340420-969b-4ccb-9f9a-7a833d8f0c43/5c340420-969b-4ccb-9f9a-7a833d8f0c43.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 714.399655] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1443f008-08cc-4566-9f24-7b224f576eab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.410218] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 714.410218] env[61964]: value = "task-1040703" [ 714.410218] env[61964]: _type = "Task" [ 714.410218] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.415925] env[61964]: DEBUG nova.compute.manager [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 714.416210] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 714.417449] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440f823b-afbf-46c7-9442-236aa8a7a30e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.443740] env[61964]: DEBUG oslo_concurrency.lockutils [req-85c76965-69f5-4807-a64b-6b26828484c3 req-1237015a-8b81-45d6-ae97-8d19851de7ee service nova] Releasing lock "refresh_cache-5c340420-969b-4ccb-9f9a-7a833d8f0c43" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.444418] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.455405] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 714.455405] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a12e7a1f-6614-4fdb-83c0-adeeaf4e50ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.465382] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 714.465382] env[61964]: value = "task-1040704" [ 714.465382] env[61964]: _type = "Task" [ 714.465382] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.482069] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040704, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.559882] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddfed24-0ff4-4b11-8efe-631a0af100dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.570397] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee77fc9-d5ae-4007-9132-223ab529442f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.606219] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ecb0a0-590d-4be0-ba11-416821bdd91f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.615291] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad058d49-d732-4e74-829e-ef07b5d77b66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.635574] env[61964]: DEBUG nova.compute.provider_tree [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.788479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.788805] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance network_info: |[{"id": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "address": "fa:16:3e:19:59:60", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d1d176f-64", "ovs_interfaceid": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 714.789294] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:59:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d1d176f-6488-4095-b0a2-52d2f6d99245', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.797344] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating folder: Project (2ccf1fe5003e484fa73ce5c32767479f). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 714.797699] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cf4cc62-e8cd-4e5c-8780-954919e746ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.816557] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created folder: Project (2ccf1fe5003e484fa73ce5c32767479f) in parent group-v230360. [ 714.817320] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating folder: Instances. Parent ref: group-v230418. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 714.817320] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b344588-f340-49ee-b6c7-79ef8a359ee0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.820712] env[61964]: DEBUG oslo_concurrency.lockutils [req-66f9d069-6dc0-4e3c-baa9-bb8fdac143ea req-90538582-5c87-4b34-b772-473c0d90c8ff service nova] Releasing lock "refresh_cache-af413b1d-9e7e-43d0-a39a-fb9dda9cf281" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.834870] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created folder: Instances in parent group-v230418. [ 714.839451] env[61964]: DEBUG oslo.service.loopingcall [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.839451] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 714.839451] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42496ea0-a989-4638-abb3-c87374b7490f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.864425] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.864425] env[61964]: value = "task-1040707" [ 714.864425] env[61964]: _type = "Task" [ 714.864425] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.872341] env[61964]: DEBUG oslo_vmware.api [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040702, 'name': PowerOnVM_Task, 'duration_secs': 0.51841} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.873279] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 714.876920] env[61964]: DEBUG nova.compute.manager [None req-9bae6ad8-43c1-4388-a244-8bb63dbe9429 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 714.880876] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e453f1c-5d10-44d8-94fc-6ce93385e6bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.883645] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040707, 'name': CreateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.921889] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040703, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.983488] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040704, 'name': PowerOffVM_Task, 'duration_secs': 0.283605} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.983488] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 714.983488] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 714.983781] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b2875b9-b8c3-4e5d-bb2b-295e791b1911 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.046225] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 715.065269] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 715.065515] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 715.065788] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Deleting the datastore file [datastore2] a2a25787-1426-4c09-892d-395e1becccc9 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 715.066093] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d160e935-3652-46ea-9621-2bcd2f9938e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.074111] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for the task: (returnval){ [ 715.074111] env[61964]: value = "task-1040709" [ 715.074111] env[61964]: _type = "Task" [ 715.074111] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.079286] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 715.079534] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.079694] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 715.079880] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.080038] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 715.080195] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 715.080399] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 715.080560] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 715.080729] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 715.080894] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 715.081092] env[61964]: DEBUG nova.virt.hardware [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 715.081875] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44485b6-6ab1-4120-bfe6-f925ddbe9d03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.092664] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d3bfda-a8fd-461e-a3d0-d5bdf9ca4e24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.096697] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.140773] env[61964]: DEBUG nova.scheduler.client.report [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.150547] env[61964]: DEBUG nova.compute.manager [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Received event network-vif-plugged-6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 715.150868] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Acquiring lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.151125] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.151430] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.151564] env[61964]: DEBUG nova.compute.manager [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] No waiting events found dispatching network-vif-plugged-6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 715.151663] env[61964]: WARNING nova.compute.manager [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Received unexpected event network-vif-plugged-6d1d176f-6488-4095-b0a2-52d2f6d99245 for instance with vm_state building and task_state spawning. [ 715.151837] env[61964]: DEBUG nova.compute.manager [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Received event network-changed-6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 715.151936] env[61964]: DEBUG nova.compute.manager [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Refreshing instance network info cache due to event network-changed-6d1d176f-6488-4095-b0a2-52d2f6d99245. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 715.152196] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Acquiring lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.152406] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Acquired lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.152605] env[61964]: DEBUG nova.network.neutron [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Refreshing network info cache for port 6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.382015] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040707, 'name': CreateVM_Task, 'duration_secs': 0.488558} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.382235] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 715.382951] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.383142] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.383485] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.383757] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebcdb738-cabe-4ffa-a737-dd6ffbc75c87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.389243] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 715.389243] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527392e8-5815-db21-1ed7-4761f7602a14" [ 715.389243] env[61964]: _type = "Task" [ 715.389243] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.400300] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527392e8-5815-db21-1ed7-4761f7602a14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.422791] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634997} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.423009] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 5c340420-969b-4ccb-9f9a-7a833d8f0c43/5c340420-969b-4ccb-9f9a-7a833d8f0c43.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 715.423289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 715.423544] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30ad10dc-0e4a-416a-9064-2a4745ab5406 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.430725] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 715.430725] env[61964]: value = "task-1040710" [ 715.430725] env[61964]: _type = "Task" [ 715.430725] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.439411] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040710, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.585232] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.646959] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.643s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.646959] env[61964]: INFO nova.compute.manager [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Migrating [ 715.646959] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.647096] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.648490] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.653018] env[61964]: INFO nova.compute.claims [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.901662] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527392e8-5815-db21-1ed7-4761f7602a14, 'name': SearchDatastore_Task, 'duration_secs': 0.022055} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.901968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.902262] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.902512] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.902664] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.902844] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.903344] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39138afe-78fc-4e33-b178-d361cec5ccfa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.918972] env[61964]: DEBUG nova.network.neutron [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Updated VIF entry in instance network info cache for port 6d1d176f-6488-4095-b0a2-52d2f6d99245. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 715.922279] env[61964]: DEBUG nova.network.neutron [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Updating instance_info_cache with network_info: [{"id": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "address": "fa:16:3e:19:59:60", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d1d176f-64", "ovs_interfaceid": "6d1d176f-6488-4095-b0a2-52d2f6d99245", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.922279] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.922457] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 715.922737] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d353bb7-c289-44fc-b93a-85ef6d039924 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.928288] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 715.928288] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52332241-5a59-cf40-7f8e-ede1e9e3f163" [ 715.928288] env[61964]: _type = "Task" [ 715.928288] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.942595] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52332241-5a59-cf40-7f8e-ede1e9e3f163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.946823] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040710, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072486} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.946927] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.947884] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9cad7a-9428-4a09-845d-7aa68b3b6912 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.972569] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 5c340420-969b-4ccb-9f9a-7a833d8f0c43/5c340420-969b-4ccb-9f9a-7a833d8f0c43.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.973281] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa3bd84f-9dd7-4fd2-9692-83ec41dde9b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.988411] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Successfully updated port: 8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 715.997612] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 715.997612] env[61964]: value = "task-1040711" [ 715.997612] env[61964]: _type = "Task" [ 715.997612] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.008543] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.086319] env[61964]: DEBUG oslo_vmware.api [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Task: {'id': task-1040709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.644864} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.086567] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.087413] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 716.087413] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 716.087413] env[61964]: INFO nova.compute.manager [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Took 1.67 seconds to destroy the instance on the hypervisor. [ 716.087579] env[61964]: DEBUG oslo.service.loopingcall [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.087717] env[61964]: DEBUG nova.compute.manager [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 716.087829] env[61964]: DEBUG nova.network.neutron [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.156700] env[61964]: INFO nova.compute.rpcapi [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 716.157243] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.425285] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f5490b3-1e26-459d-996f-75258effda60 req-804236be-9651-4115-bf5f-29bc9dde4907 service nova] Releasing lock "refresh_cache-df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.441368] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52332241-5a59-cf40-7f8e-ede1e9e3f163, 'name': SearchDatastore_Task, 'duration_secs': 0.049871} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.442468] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50eefce4-9daa-49a0-bb34-8cc6523a4427 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.449912] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 716.449912] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fd1d46-e7a5-fd60-68c8-228ca1ed298b" [ 716.449912] env[61964]: _type = "Task" [ 716.449912] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.459143] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fd1d46-e7a5-fd60-68c8-228ca1ed298b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.491649] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.491839] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.492095] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.498856] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.499054] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.510351] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040711, 'name': ReconfigVM_Task, 'duration_secs': 0.290179} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.510748] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 5c340420-969b-4ccb-9f9a-7a833d8f0c43/5c340420-969b-4ccb-9f9a-7a833d8f0c43.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 716.511479] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2257ae55-d4e9-4942-90bc-dad1184be407 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.520104] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 716.520104] env[61964]: value = "task-1040712" [ 716.520104] env[61964]: _type = "Task" [ 716.520104] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.529537] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040712, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.677036] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.677036] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.677515] env[61964]: DEBUG nova.network.neutron [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.917401] env[61964]: DEBUG nova.network.neutron [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.960580] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fd1d46-e7a5-fd60-68c8-228ca1ed298b, 'name': SearchDatastore_Task, 'duration_secs': 0.013258} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.963416] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.963464] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.964515] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d57cbd1-ac14-45b4-a6ca-a6caf32be8ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.972017] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 716.972017] env[61964]: value = "task-1040713" [ 716.972017] env[61964]: _type = "Task" [ 716.972017] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.983352] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040713, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.027286] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.036604] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040712, 'name': Rename_Task, 'duration_secs': 0.150348} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.037226] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 717.037517] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-799a2821-2485-4588-b598-eb2dd64014cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.046068] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 717.046068] env[61964]: value = "task-1040714" [ 717.046068] env[61964]: _type = "Task" [ 717.046068] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.051427] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce7af60-b1d4-4027-987a-48a39291fb28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.064196] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.067526] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8fae6b-07a3-4d90-99e4-58c1309c05bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.108374] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a35c0d8-7690-44a0-9bd1-067a43060ccb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.116592] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209f0cf0-efda-440d-b45f-8de5cdb07478 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.132732] env[61964]: DEBUG nova.compute.provider_tree [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.276506] env[61964]: DEBUG nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Received event network-vif-plugged-8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 717.276820] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.277017] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.277206] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.277374] env[61964]: DEBUG nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] No waiting events found dispatching network-vif-plugged-8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 717.277549] env[61964]: WARNING nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Received unexpected event network-vif-plugged-8b2c44a6-df10-43fa-9afa-85167edbb650 for instance with vm_state building and task_state spawning. [ 717.277717] env[61964]: DEBUG nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Received event network-changed-8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 717.277879] env[61964]: DEBUG nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Refreshing instance network info cache due to event network-changed-8b2c44a6-df10-43fa-9afa-85167edbb650. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 717.278064] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.285607] env[61964]: DEBUG nova.network.neutron [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.421294] env[61964]: INFO nova.compute.manager [-] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Took 1.33 seconds to deallocate network for instance. [ 717.488481] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040713, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.501945] env[61964]: DEBUG nova.network.neutron [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.560271] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040714, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.640322] env[61964]: DEBUG nova.scheduler.client.report [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.788279] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.788600] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Instance network_info: |[{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 717.788893] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.789096] env[61964]: DEBUG nova.network.neutron [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Refreshing network info cache for port 8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.790425] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:a0:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b2c44a6-df10-43fa-9afa-85167edbb650', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.797783] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating folder: Project (9b627b25af0c4982a22c8c0bd28641a3). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.798891] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8a34186-db76-496a-a467-51932119c0d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.810796] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created folder: Project (9b627b25af0c4982a22c8c0bd28641a3) in parent group-v230360. [ 717.810979] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating folder: Instances. Parent ref: group-v230421. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.811227] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ecb64b1b-a3aa-444f-ba3d-df206f4b6c40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.819595] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created folder: Instances in parent group-v230421. [ 717.819824] env[61964]: DEBUG oslo.service.loopingcall [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.820016] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 717.820219] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c1b131f-d342-4318-9631-5f52acb975fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.838151] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.838151] env[61964]: value = "task-1040717" [ 717.838151] env[61964]: _type = "Task" [ 717.838151] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.845680] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040717, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.929620] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.983422] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040713, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594551} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.983720] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.983943] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.984243] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f52d8cdf-f1ee-4307-90be-3e43bc761fa3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.991912] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 717.991912] env[61964]: value = "task-1040718" [ 717.991912] env[61964]: _type = "Task" [ 717.991912] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.000286] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.005008] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.057249] env[61964]: DEBUG oslo_vmware.api [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040714, 'name': PowerOnVM_Task, 'duration_secs': 0.57825} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.059050] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 718.059050] env[61964]: INFO nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Took 8.14 seconds to spawn the instance on the hypervisor. [ 718.059050] env[61964]: DEBUG nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 718.059418] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e4e752-437e-432a-8e92-7f37568948c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.144802] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.144802] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 718.147401] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.341s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.148722] env[61964]: INFO nova.compute.claims [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.349855] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040717, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.505017] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065867} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.505017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.505017] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8f3b81-2889-45db-9d85-3a5eb52d07a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.528719] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.529956] env[61964]: DEBUG nova.network.neutron [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updated VIF entry in instance network info cache for port 8b2c44a6-df10-43fa-9afa-85167edbb650. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 718.531031] env[61964]: DEBUG nova.network.neutron [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.531522] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-231629f1-40c0-4feb-bc8d-3d884a74fb78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.556224] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 718.556224] env[61964]: value = "task-1040719" [ 718.556224] env[61964]: _type = "Task" [ 718.556224] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.567353] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040719, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.582049] env[61964]: INFO nova.compute.manager [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Took 32.42 seconds to build instance. [ 718.653469] env[61964]: DEBUG nova.compute.utils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.656698] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 718.656937] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 718.706595] env[61964]: DEBUG nova.policy [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad79740d20af45e98a2ec8c23066244c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a99659516ad4274b631674f20b36280', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.854039] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040717, 'name': CreateVM_Task, 'duration_secs': 0.927656} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.854328] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 718.858018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.858018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.858018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.858018] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48bf2546-b05e-4fec-bfde-8695a1fd919c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.862934] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 718.862934] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522bf98c-9c5c-dda9-072c-b45f95e9b96a" [ 718.862934] env[61964]: _type = "Task" [ 718.862934] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.872818] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522bf98c-9c5c-dda9-072c-b45f95e9b96a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.048515] env[61964]: DEBUG oslo_concurrency.lockutils [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.048827] env[61964]: DEBUG nova.compute.manager [req-aa2211b4-0b53-4845-b76f-47e9450d57f4 req-e06c5b8d-6532-495b-babd-80a673068ddd service nova] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Received event network-vif-deleted-6b0cb07c-91a0-4e3f-a285-82c6ee167d97 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 719.058324] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Successfully created port: ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.070990] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.081631] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.084600] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5f702a29-6c8f-4d38-be6a-cb488c3901c7 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.482s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.086102] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.086102] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.086403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.086548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.091714] env[61964]: INFO nova.compute.manager [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Terminating instance [ 719.157696] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.379065] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522bf98c-9c5c-dda9-072c-b45f95e9b96a, 'name': SearchDatastore_Task, 'duration_secs': 0.010816} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.379286] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.379560] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.379885] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.380052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.380240] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.380822] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10cc7069-6074-41de-9e02-f83714b23a7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.390411] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.390609] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.391395] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96bebf8f-f12f-4c88-9459-d24ab78a7150 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.397067] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 719.397067] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525b7a7d-7529-0a48-f6d4-79256f23893d" [ 719.397067] env[61964]: _type = "Task" [ 719.397067] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.409246] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525b7a7d-7529-0a48-f6d4-79256f23893d, 'name': SearchDatastore_Task, 'duration_secs': 0.009075} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.410028] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db6ca356-d5d1-42ba-876d-de1bf407c02d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.417867] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 719.417867] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db781c-e0fb-2890-ed8a-fbe5979761ee" [ 719.417867] env[61964]: _type = "Task" [ 719.417867] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.429392] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db781c-e0fb-2890-ed8a-fbe5979761ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.538238] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06067e3c-a7a4-487d-800a-05cc56517742 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.560652] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 719.572779] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040719, 'name': ReconfigVM_Task, 'duration_secs': 0.730697} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.573492] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Reconfigured VM instance instance-0000002b to attach disk [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.574505] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532e777d-2129-4e33-9d91-0adc1a34e860 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.576892] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-566ea216-7f2a-42e5-b0b0-b605efffc0c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.584019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ca635a-2d24-4c06-b3c7-5c06f2835b0c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.587461] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 719.587461] env[61964]: value = "task-1040720" [ 719.587461] env[61964]: _type = "Task" [ 719.587461] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.615046] env[61964]: DEBUG nova.compute.manager [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 719.615046] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 719.615046] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 719.619412] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a46dd2-88ed-46f8-ad9c-7a6235795504 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.622514] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a845a8f7-ace2-40e4-936a-256942370f7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.629183] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040720, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.635418] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 719.637410] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cc3d1e1-6a94-4c9d-8273-2d31d1f60258 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.644326] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06007166-9860-4ce4-8bea-2d1a47577bf2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.649914] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 719.649914] env[61964]: value = "task-1040721" [ 719.649914] env[61964]: _type = "Task" [ 719.649914] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.661383] env[61964]: DEBUG nova.compute.provider_tree [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.667309] env[61964]: INFO nova.virt.block_device [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Booting with volume 8352b6f4-47b3-40e9-93a1-3ce98bf146a9 at /dev/sda [ 719.669051] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.716729] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5bf5e47-c070-42e9-8c43-ca14d4b03212 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.730373] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3493a5c2-c0e9-40f6-855f-89669447d9a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.771911] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aeb1ea47-19ea-43b9-9e34-ff40e6d8e661 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.781060] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11e4df2-54b1-4dff-94c8-2cc65919937b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.812439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2bbe52-877d-4845-8961-c3d34c6c49ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.820763] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788b0f32-2432-48bc-8da6-1805793455d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.835940] env[61964]: DEBUG nova.virt.block_device [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating existing volume attachment record: b73d893b-9964-40b7-bd6f-6a9b0d438851 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 719.928297] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db781c-e0fb-2890-ed8a-fbe5979761ee, 'name': SearchDatastore_Task, 'duration_secs': 0.009155} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.928635] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.928901] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 719.929186] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9caa38d4-37a0-4a86-bf17-35fedd66ce33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.936834] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 719.936834] env[61964]: value = "task-1040722" [ 719.936834] env[61964]: _type = "Task" [ 719.936834] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.945804] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040722, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.068982] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 720.069330] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b2b6fa2-c735-4592-a1cf-e844c99df11a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.078297] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 720.078297] env[61964]: value = "task-1040723" [ 720.078297] env[61964]: _type = "Task" [ 720.078297] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.089499] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040723, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.100663] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040720, 'name': Rename_Task, 'duration_secs': 0.187279} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.100976] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.101344] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1aeadc69-4781-43ac-b9d4-668fd382e60f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.112660] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 720.112660] env[61964]: value = "task-1040724" [ 720.112660] env[61964]: _type = "Task" [ 720.112660] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.122563] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.142851] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.164062] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040721, 'name': PowerOffVM_Task, 'duration_secs': 0.232092} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.164512] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 720.164708] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 720.165009] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13ff06d8-4061-44d8-9d87-e6f208c571aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.171282] env[61964]: DEBUG nova.scheduler.client.report [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.249886] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 720.250230] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 720.250474] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Deleting the datastore file [datastore1] 5c340420-969b-4ccb-9f9a-7a833d8f0c43 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 720.250784] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8830564c-4902-4dcd-bc6f-0bf20f41a7a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.262188] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for the task: (returnval){ [ 720.262188] env[61964]: value = "task-1040726" [ 720.262188] env[61964]: _type = "Task" [ 720.262188] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.278362] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.448665] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040722, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.593012] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040723, 'name': PowerOffVM_Task, 'duration_secs': 0.48456} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.593346] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 720.593539] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 720.624521] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040724, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.679025] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.679025] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 720.680841] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.453s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.683068] env[61964]: INFO nova.compute.claims [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.690079] env[61964]: DEBUG nova.compute.manager [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Received event network-vif-plugged-ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 720.692171] env[61964]: DEBUG oslo_concurrency.lockutils [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] Acquiring lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.692171] env[61964]: DEBUG oslo_concurrency.lockutils [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.692171] env[61964]: DEBUG oslo_concurrency.lockutils [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.692171] env[61964]: DEBUG nova.compute.manager [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] No waiting events found dispatching network-vif-plugged-ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 720.692442] env[61964]: WARNING nova.compute.manager [req-5021c030-9587-4681-9317-f161ee46ac7f req-fcac9177-6b79-471a-95be-9a9522f302c1 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Received unexpected event network-vif-plugged-ef8e5acc-8a42-44d0-84c4-57803733186b for instance with vm_state building and task_state block_device_mapping. [ 720.776566] env[61964]: DEBUG oslo_vmware.api [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Task: {'id': task-1040726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284418} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.776723] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 720.776984] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 720.777236] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 720.777449] env[61964]: INFO nova.compute.manager [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Took 1.16 seconds to destroy the instance on the hypervisor. [ 720.777731] env[61964]: DEBUG oslo.service.loopingcall [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.777953] env[61964]: DEBUG nova.compute.manager [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 720.778224] env[61964]: DEBUG nova.network.neutron [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.804951] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Successfully updated port: ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 720.949195] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040722, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521559} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.949701] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 720.949915] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.950221] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4dcc4575-bf17-4e40-8572-2c8b2e996137 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.961890] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 720.961890] env[61964]: value = "task-1040727" [ 720.961890] env[61964]: _type = "Task" [ 720.961890] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.972552] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040727, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.104695] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 721.105327] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.105327] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 721.105606] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.105800] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 721.106091] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 721.106471] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 721.106765] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 721.107191] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 721.107501] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 721.112022] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 721.119180] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37898590-d3e2-49b6-a9e7-c1c920ddc127 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.132621] env[61964]: DEBUG nova.compute.manager [req-13050dee-239d-47f2-beeb-090824bd92e5 req-b684206d-45e4-4080-ac35-179f6fef28f8 service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Received event network-vif-deleted-bdaf1e74-2420-4ca8-9dfe-ba0b13624e15 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 721.133034] env[61964]: INFO nova.compute.manager [req-13050dee-239d-47f2-beeb-090824bd92e5 req-b684206d-45e4-4080-ac35-179f6fef28f8 service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Neutron deleted interface bdaf1e74-2420-4ca8-9dfe-ba0b13624e15; detaching it from the instance and deleting it from the info cache [ 721.133341] env[61964]: DEBUG nova.network.neutron [req-13050dee-239d-47f2-beeb-090824bd92e5 req-b684206d-45e4-4080-ac35-179f6fef28f8 service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.145383] env[61964]: DEBUG oslo_vmware.api [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040724, 'name': PowerOnVM_Task, 'duration_secs': 0.540309} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.147681] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.148130] env[61964]: INFO nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Took 8.67 seconds to spawn the instance on the hypervisor. [ 721.148468] env[61964]: DEBUG nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 721.149198] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 721.149198] env[61964]: value = "task-1040728" [ 721.149198] env[61964]: _type = "Task" [ 721.149198] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.150156] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a29d39-5e41-4fcd-8a17-9a61ab1aa0fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.168490] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040728, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.187925] env[61964]: DEBUG nova.compute.utils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.189373] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 721.189574] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.243462] env[61964]: DEBUG nova.policy [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.311403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.311681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquired lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.311855] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.473229] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040727, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.179183} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.473229] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.474234] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09286677-bb5b-4cc2-b996-8dba3a47bbd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.500275] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.501723] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08846a3b-709e-4044-8ecc-f3fa042cbd8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.526236] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 721.526236] env[61964]: value = "task-1040729" [ 721.526236] env[61964]: _type = "Task" [ 721.526236] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.534503] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040729, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.577792] env[61964]: DEBUG nova.network.neutron [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.639638] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b21578c-7336-492b-b664-47faa2c9fec8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.651020] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6e2f9a-3e36-4285-80f1-a86e18090691 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.680133] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040728, 'name': ReconfigVM_Task, 'duration_secs': 0.162019} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.689464] env[61964]: INFO nova.compute.manager [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Took 33.39 seconds to build instance. [ 721.691227] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 721.695056] env[61964]: DEBUG nova.compute.manager [req-13050dee-239d-47f2-beeb-090824bd92e5 req-b684206d-45e4-4080-ac35-179f6fef28f8 service nova] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Detach interface failed, port_id=bdaf1e74-2420-4ca8-9dfe-ba0b13624e15, reason: Instance 5c340420-969b-4ccb-9f9a-7a833d8f0c43 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 721.698738] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 721.781076] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Successfully created port: 33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.868173] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.956268] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 721.956994] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 721.957234] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.957400] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 721.957586] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.957735] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 721.957885] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 721.958564] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 721.958849] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 721.959098] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 721.959539] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 721.959748] env[61964]: DEBUG nova.virt.hardware [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 721.960962] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf308f0-6c96-49a4-999b-aea19e818ac1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.972723] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3381c9-5899-49bd-a5e1-58433808acdd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.042016] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040729, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.082944] env[61964]: INFO nova.compute.manager [-] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Took 1.30 seconds to deallocate network for instance. [ 722.090890] env[61964]: DEBUG nova.network.neutron [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating instance_info_cache with network_info: [{"id": "ef8e5acc-8a42-44d0-84c4-57803733186b", "address": "fa:16:3e:2e:93:49", "network": {"id": "c3e10648-bb91-4e2c-9f64-9b0479002071", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-192055358-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a99659516ad4274b631674f20b36280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e5acc-8a", "ovs_interfaceid": "ef8e5acc-8a42-44d0-84c4-57803733186b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.191474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e7da8b6-ff54-4047-9859-18234e62667a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.908s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.197281] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:21:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='4d7c58f5-7642-43b5-a3cb-6aff98a75b9e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-994792262',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 722.197563] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.197744] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 722.198126] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.198185] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 722.198349] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 722.198576] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 722.198752] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 722.198923] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 722.199158] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 722.199352] env[61964]: DEBUG nova.virt.hardware [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 722.205049] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfiguring VM instance instance-00000022 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 722.212281] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba4a9415-01c4-4a96-9035-d32142f05e6b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.234269] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 722.234269] env[61964]: value = "task-1040730" [ 722.234269] env[61964]: _type = "Task" [ 722.234269] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.239473] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397387e3-44e7-4536-94fa-7e7c781b0c10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.246049] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040730, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.250849] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6a6012-12d7-4474-871e-ba6ca4bd6af2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.286528] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b26dc5-e94c-423d-a68e-323f3364118a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.294836] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd19f253-1c5d-4204-8dcb-d983add70829 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.310557] env[61964]: DEBUG nova.compute.provider_tree [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.534966] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040729, 'name': ReconfigVM_Task, 'duration_secs': 0.603061} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.535279] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfigured VM instance instance-0000002c to attach disk [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.535907] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-109b31ef-acea-421a-9f41-9357939c340c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.542806] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 722.542806] env[61964]: value = "task-1040731" [ 722.542806] env[61964]: _type = "Task" [ 722.542806] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.551370] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040731, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.594262] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.594774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Releasing lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.595104] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Instance network_info: |[{"id": "ef8e5acc-8a42-44d0-84c4-57803733186b", "address": "fa:16:3e:2e:93:49", "network": {"id": "c3e10648-bb91-4e2c-9f64-9b0479002071", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-192055358-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a99659516ad4274b631674f20b36280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e5acc-8a", "ovs_interfaceid": "ef8e5acc-8a42-44d0-84c4-57803733186b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 722.595494] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:93:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef8e5acc-8a42-44d0-84c4-57803733186b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.602826] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Creating folder: Project (6a99659516ad4274b631674f20b36280). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 722.603083] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ccf623c7-e261-45a5-855b-082dc2494873 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.617253] env[61964]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 722.617410] env[61964]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61964) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 722.617917] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Folder already exists: Project (6a99659516ad4274b631674f20b36280). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.617917] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Creating folder: Instances. Parent ref: group-v230370. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 722.618438] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9548aae8-cd2b-480a-af66-876abe29173e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.627770] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Created folder: Instances in parent group-v230370. [ 722.627998] env[61964]: DEBUG oslo.service.loopingcall [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.628199] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 722.628393] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-559cbbb7-5516-4137-9cad-2818bd1c4f56 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.646731] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.646731] env[61964]: value = "task-1040734" [ 722.646731] env[61964]: _type = "Task" [ 722.646731] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.654383] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040734, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.694373] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 722.713990] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 722.724081] env[61964]: DEBUG nova.compute.manager [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Received event network-changed-ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 722.724321] env[61964]: DEBUG nova.compute.manager [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Refreshing instance network info cache due to event network-changed-ef8e5acc-8a42-44d0-84c4-57803733186b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 722.724564] env[61964]: DEBUG oslo_concurrency.lockutils [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] Acquiring lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.724976] env[61964]: DEBUG oslo_concurrency.lockutils [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] Acquired lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.724976] env[61964]: DEBUG nova.network.neutron [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Refreshing network info cache for port ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 722.741518] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 722.741743] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.741985] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 722.742186] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.742376] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 722.742555] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 722.742794] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 722.742961] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 722.743154] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 722.743371] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 722.743595] env[61964]: DEBUG nova.virt.hardware [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 722.744962] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45452bd-a279-45f3-88af-e2f3a9445900 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.754029] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040730, 'name': ReconfigVM_Task, 'duration_secs': 0.191107} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.756219] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfigured VM instance instance-00000022 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 722.757040] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14977c9e-5b0d-47ed-b96b-bf8af53b55e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.760424] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880f77c0-3550-47dc-9b01-8c16fbce977d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.790987] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.799096] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffdf0eee-517c-4761-a1e0-331a6b4f8b59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.820795] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 722.820795] env[61964]: value = "task-1040735" [ 722.820795] env[61964]: _type = "Task" [ 722.820795] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.830118] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.831297] env[61964]: ERROR nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [req-a7955596-b51a-45c5-b139-44955a33fa39] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a7955596-b51a-45c5-b139-44955a33fa39"}]} [ 722.853469] env[61964]: DEBUG nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 722.868494] env[61964]: DEBUG nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 722.868776] env[61964]: DEBUG nova.compute.provider_tree [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.884153] env[61964]: DEBUG nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 722.906925] env[61964]: DEBUG nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 722.977157] env[61964]: INFO nova.compute.manager [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Rebuilding instance [ 723.019235] env[61964]: DEBUG nova.compute.manager [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 723.020187] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01162542-5060-4713-8e36-173b614e61a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.053741] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040731, 'name': Rename_Task, 'duration_secs': 0.275287} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.057325] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 723.058029] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac355f87-0b01-4ed2-b2d1-fe9cc6dccfdf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.067741] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 723.067741] env[61964]: value = "task-1040736" [ 723.067741] env[61964]: _type = "Task" [ 723.067741] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.079789] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040736, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.157292] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040734, 'name': CreateVM_Task, 'duration_secs': 0.360515} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.159894] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 723.160743] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sda', 'delete_on_termination': True, 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': 'b73d893b-9964-40b7-bd6f-6a9b0d438851', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230373', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'name': 'volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6ee845c3-5f79-4704-8b7b-cd3770202647', 'attached_at': '', 'detached_at': '', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'serial': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9'}, 'volume_type': None}], 'swap': None} {{(pid=61964) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 723.161099] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Root volume attach. Driver type: vmdk {{(pid=61964) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 723.161816] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e122ba4d-7876-4362-9e72-ec1ea6e3dafe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.170456] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb82de3-59ca-4004-854d-09943eaeeeb9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.180561] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571a9786-c43f-485a-b062-8614ec8920eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.190549] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-4f0f09f8-644d-49f3-9be4-e2a72cef3704 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.200921] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 723.200921] env[61964]: value = "task-1040737" [ 723.200921] env[61964]: _type = "Task" [ 723.200921] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.214953] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040737, 'name': RelocateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.215956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.332987] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040735, 'name': ReconfigVM_Task, 'duration_secs': 0.374697} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.333892] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.333892] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 723.379244] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745521b8-b530-464c-9a68-eab35b5aa15d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.388925] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ec4f84-3165-409d-92e3-e30a3c2480f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.400170] env[61964]: DEBUG nova.compute.manager [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Received event network-vif-plugged-33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 723.400399] env[61964]: DEBUG oslo_concurrency.lockutils [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] Acquiring lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.400612] env[61964]: DEBUG oslo_concurrency.lockutils [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.400802] env[61964]: DEBUG oslo_concurrency.lockutils [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.400962] env[61964]: DEBUG nova.compute.manager [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] No waiting events found dispatching network-vif-plugged-33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 723.401203] env[61964]: WARNING nova.compute.manager [req-db4c1e7b-1a4f-40f0-aff5-9585d2f19b64 req-23941ca7-26e8-47e3-8a32-c245f917c784 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Received unexpected event network-vif-plugged-33b1704a-2711-4fb6-bae6-94dc5716b4d3 for instance with vm_state building and task_state spawning. [ 723.434693] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a3c539-d7c7-48fa-ac17-7f957ff2b48d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.444942] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923edba1-579c-4075-b3cc-335b62e360c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.461585] env[61964]: DEBUG nova.compute.provider_tree [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.478990] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Successfully updated port: 33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.511146] env[61964]: DEBUG nova.network.neutron [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updated VIF entry in instance network info cache for port ef8e5acc-8a42-44d0-84c4-57803733186b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 723.511533] env[61964]: DEBUG nova.network.neutron [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating instance_info_cache with network_info: [{"id": "ef8e5acc-8a42-44d0-84c4-57803733186b", "address": "fa:16:3e:2e:93:49", "network": {"id": "c3e10648-bb91-4e2c-9f64-9b0479002071", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-192055358-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a99659516ad4274b631674f20b36280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e5acc-8a", "ovs_interfaceid": "ef8e5acc-8a42-44d0-84c4-57803733186b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.581270] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040736, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.715211] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040737, 'name': RelocateVM_Task, 'duration_secs': 0.472838} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.715550] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 723.715759] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230373', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'name': 'volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6ee845c3-5f79-4704-8b7b-cd3770202647', 'attached_at': '', 'detached_at': '', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'serial': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 723.716544] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4a6752-6780-4e89-8b90-780e4d77e3f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.733026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d34c2d-98c3-40cb-a1e8-c6dae48b7ffa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.756086] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9/volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.756349] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c0ffe7c-dd35-4110-a16f-703c900bd11f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.776414] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 723.776414] env[61964]: value = "task-1040738" [ 723.776414] env[61964]: _type = "Task" [ 723.776414] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.785037] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040738, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.843300] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c752f35d-defb-4965-8d08-d38758d6c3e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.864603] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edcaa34-72ff-455d-ae7e-72beafc693f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.884485] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 723.982080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.982348] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.983536] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.008120] env[61964]: DEBUG nova.scheduler.client.report [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 724.008421] env[61964]: DEBUG nova.compute.provider_tree [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 69 to 70 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 724.008605] env[61964]: DEBUG nova.compute.provider_tree [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.014698] env[61964]: DEBUG oslo_concurrency.lockutils [req-3fe9dff9-1c9d-49e2-8c33-4f53ca3a9c56 req-a6ed7123-cb29-4851-9cfd-00bba4cb5e07 service nova] Releasing lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.035544] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 724.036155] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-431c669c-b97b-4dd6-a583-a3937acaba81 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.045345] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 724.045345] env[61964]: value = "task-1040739" [ 724.045345] env[61964]: _type = "Task" [ 724.045345] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.057205] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.079124] env[61964]: DEBUG oslo_vmware.api [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040736, 'name': PowerOnVM_Task, 'duration_secs': 0.538919} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.079393] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 724.079601] env[61964]: INFO nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 9.03 seconds to spawn the instance on the hypervisor. [ 724.079782] env[61964]: DEBUG nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 724.080577] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0666804d-767d-4560-b9d6-45be394bdfd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.286985] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040738, 'name': ReconfigVM_Task, 'duration_secs': 0.268688} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.287293] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Reconfigured VM instance instance-0000002d to attach disk [datastore1] volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9/volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.291922] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12db9d36-9d33-4553-b195-a47198831b3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.309049] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 724.309049] env[61964]: value = "task-1040740" [ 724.309049] env[61964]: _type = "Task" [ 724.309049] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.316696] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040740, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.443538] env[61964]: DEBUG nova.network.neutron [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Port 03f523b6-4fc2-4fe2-a751-d33c00f6849f binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 724.514527] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.833s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.515093] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 724.519389] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.521695] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.184s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.523210] env[61964]: INFO nova.compute.claims [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.557878] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040739, 'name': PowerOffVM_Task, 'duration_secs': 0.434364} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.558874] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.559127] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 724.559909] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc76fe21-23cf-43c3-94d2-ca2dc022e467 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.571290] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.571560] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5a037ba-4035-4b0e-828b-6510d7b30e28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.604186] env[61964]: INFO nova.compute.manager [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 33.47 seconds to build instance. [ 724.667985] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.667985] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.667985] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.667985] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff8f96a9-051a-464f-abe3-553010ac0a96 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.675596] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 724.675596] env[61964]: value = "task-1040742" [ 724.675596] env[61964]: _type = "Task" [ 724.675596] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.684693] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040742, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.716190] env[61964]: DEBUG nova.network.neutron [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Updating instance_info_cache with network_info: [{"id": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "address": "fa:16:3e:ba:5d:77", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33b1704a-27", "ovs_interfaceid": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.818967] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040740, 'name': ReconfigVM_Task, 'duration_secs': 0.172942} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.818967] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230373', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'name': 'volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6ee845c3-5f79-4704-8b7b-cd3770202647', 'attached_at': '', 'detached_at': '', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'serial': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 724.819240] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aeca5fe2-ba8c-4d8c-b39b-a22f260b83db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.826320] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 724.826320] env[61964]: value = "task-1040743" [ 724.826320] env[61964]: _type = "Task" [ 724.826320] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.835337] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040743, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.028258] env[61964]: DEBUG nova.compute.utils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.032466] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 725.032640] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.108388] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d83c9d55-036b-45d8-9415-52589304f1a8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.800s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.186736] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040742, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.415933} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.187019] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.187225] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 725.187404] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 725.219232] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.219565] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Instance network_info: |[{"id": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "address": "fa:16:3e:ba:5d:77", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33b1704a-27", "ovs_interfaceid": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 725.220060] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:5d:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33b1704a-2711-4fb6-bae6-94dc5716b4d3', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.234580] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating folder: Project (fb92e5b578d54f1499b00aa08e7841c2). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 725.239595] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-466fcedb-1081-4175-829b-b270eba179e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.253023] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created folder: Project (fb92e5b578d54f1499b00aa08e7841c2) in parent group-v230360. [ 725.253262] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating folder: Instances. Parent ref: group-v230426. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 725.253511] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e110bf30-db48-467e-91d1-ac170d7afd11 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.264957] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created folder: Instances in parent group-v230426. [ 725.265253] env[61964]: DEBUG oslo.service.loopingcall [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.265464] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 725.265681] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da4ebd22-39d7-438e-966e-439b84be54eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.288430] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.288430] env[61964]: value = "task-1040746" [ 725.288430] env[61964]: _type = "Task" [ 725.288430] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.298519] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040746, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.301700] env[61964]: DEBUG nova.policy [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04b4d3d3917f4adf908d2396a352824f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da34d9d49bcc4fab81bbeb4cdf4e3992', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.337831] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040743, 'name': Rename_Task, 'duration_secs': 0.258112} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.338133] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 725.338389] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-309d8ff2-217f-4df0-91e8-b9e23586a7a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.346288] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 725.346288] env[61964]: value = "task-1040747" [ 725.346288] env[61964]: _type = "Task" [ 725.346288] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.359753] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.469318] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "65566c20-6d69-471c-b098-3c30c01d9955-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.469697] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.470113] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.532917] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 725.581207] env[61964]: DEBUG nova.compute.manager [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Received event network-changed-33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 725.581207] env[61964]: DEBUG nova.compute.manager [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Refreshing instance network info cache due to event network-changed-33b1704a-2711-4fb6-bae6-94dc5716b4d3. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 725.581207] env[61964]: DEBUG oslo_concurrency.lockutils [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] Acquiring lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.581207] env[61964]: DEBUG oslo_concurrency.lockutils [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] Acquired lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.581207] env[61964]: DEBUG nova.network.neutron [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Refreshing network info cache for port 33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.608215] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 725.608215] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 725.615820] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 725.675183] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Successfully created port: c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.799662] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040746, 'name': CreateVM_Task, 'duration_secs': 0.416669} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.799844] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 725.802080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.802080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.802080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 725.802080] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73d5acc3-3099-4f72-b9cc-733101e4f8ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.810980] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 725.810980] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52708602-c6f3-cea7-bdf7-2f971aeabaa9" [ 725.810980] env[61964]: _type = "Task" [ 725.810980] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.823781] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52708602-c6f3-cea7-bdf7-2f971aeabaa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.863105] env[61964]: DEBUG oslo_vmware.api [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040747, 'name': PowerOnVM_Task, 'duration_secs': 0.48208} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.863105] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.863105] env[61964]: INFO nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Took 3.91 seconds to spawn the instance on the hypervisor. [ 725.863105] env[61964]: DEBUG nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 725.863744] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f894b1-3271-4773-9971-2d9cbf766352 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.077977] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f22a9c-865d-4e38-b848-eed8341535ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.088785] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284dc147-aa47-421d-bf3d-4dc9dd0c23ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.137613] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e41f1ed-1be6-4933-b146-16f757370759 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.140421] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 726.141599] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 726.141599] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 726.156225] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7baae4-a2f4-4afe-9f1d-579c2932f448 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.166644] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.182866] env[61964]: DEBUG nova.compute.provider_tree [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.237763] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 726.237885] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.238081] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.238352] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.238575] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.238806] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 726.239200] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 726.239300] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 726.239549] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 726.240265] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 726.240265] env[61964]: DEBUG nova.virt.hardware [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 726.241391] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e7758b-5180-448a-a791-35999c25174a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.250695] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d0d3ba-3871-427d-aee7-f0af32981420 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.272079] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:59:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d1d176f-6488-4095-b0a2-52d2f6d99245', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.279798] env[61964]: DEBUG oslo.service.loopingcall [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.280072] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 726.280305] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4963f67-e75c-4d23-aef8-85043f222197 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.304470] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.304470] env[61964]: value = "task-1040748" [ 726.304470] env[61964]: _type = "Task" [ 726.304470] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.313653] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040748, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.322766] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52708602-c6f3-cea7-bdf7-2f971aeabaa9, 'name': SearchDatastore_Task, 'duration_secs': 0.015598} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.323074] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.323321] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.323564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.323718] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.323901] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.324191] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c491e296-2b95-4999-aa69-7c8d1b964554 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.336338] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.336568] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.337382] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc02117c-0e21-4b92-b06c-e176821a0ae7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.343841] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 726.343841] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bf019d-f13f-1676-7b54-7907e92c654d" [ 726.343841] env[61964]: _type = "Task" [ 726.343841] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.358148] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bf019d-f13f-1676-7b54-7907e92c654d, 'name': SearchDatastore_Task, 'duration_secs': 0.011675} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.358961] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a545154-2ddf-46e8-9ffe-0967b16c1ab1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.366533] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 726.366533] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6cc99-9a4f-7426-b17a-e81d755118ee" [ 726.366533] env[61964]: _type = "Task" [ 726.366533] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.375864] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6cc99-9a4f-7426-b17a-e81d755118ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.394503] env[61964]: INFO nova.compute.manager [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Took 33.78 seconds to build instance. [ 726.424725] env[61964]: DEBUG nova.network.neutron [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Updated VIF entry in instance network info cache for port 33b1704a-2711-4fb6-bae6-94dc5716b4d3. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.425179] env[61964]: DEBUG nova.network.neutron [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Updating instance_info_cache with network_info: [{"id": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "address": "fa:16:3e:ba:5d:77", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33b1704a-27", "ovs_interfaceid": "33b1704a-2711-4fb6-bae6-94dc5716b4d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.545456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.545681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.545859] env[61964]: DEBUG nova.network.neutron [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.547912] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 726.576554] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 726.576554] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.576554] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.576915] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.576915] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.576915] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 726.576915] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 726.576915] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 726.577173] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 726.577173] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 726.577173] env[61964]: DEBUG nova.virt.hardware [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 726.578509] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c365c3b-67bb-4b26-92a2-b665b785801d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.589717] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef75c04f-c9d6-41ef-b69b-e7e2a1de1ba7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.647070] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 726.647281] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 726.647415] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 726.688237] env[61964]: DEBUG nova.scheduler.client.report [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.696019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.696019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.696019] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 726.696019] env[61964]: DEBUG nova.objects.instance [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lazy-loading 'info_cache' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.816052] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040748, 'name': CreateVM_Task, 'duration_secs': 0.453259} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.816052] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.816197] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.816391] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.816779] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.817085] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f286808d-b4af-47bd-a9c8-3b129d128efd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.823236] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 726.823236] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6038e-ccef-9d96-2f17-4ddcfe19b368" [ 726.823236] env[61964]: _type = "Task" [ 726.823236] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.832148] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6038e-ccef-9d96-2f17-4ddcfe19b368, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.875687] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6cc99-9a4f-7426-b17a-e81d755118ee, 'name': SearchDatastore_Task, 'duration_secs': 0.010871} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.875969] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.876246] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0ea19f52-b700-405f-8c3c-fbaa3e17b36a/0ea19f52-b700-405f-8c3c-fbaa3e17b36a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 726.876529] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55da1d5a-10dc-44b4-a82d-891947b7cf02 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.885222] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 726.885222] env[61964]: value = "task-1040749" [ 726.885222] env[61964]: _type = "Task" [ 726.885222] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.896627] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.896627] env[61964]: DEBUG oslo_concurrency.lockutils [None req-becbae3a-bd5b-4a27-a05c-8036776402b2 tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.602s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.928201] env[61964]: DEBUG oslo_concurrency.lockutils [req-09fc1073-783d-474d-adcf-bf2ca400aa9c req-7190cb62-df80-4c00-8212-212c68b04b09 service nova] Releasing lock "refresh_cache-0ea19f52-b700-405f-8c3c-fbaa3e17b36a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.083565] env[61964]: DEBUG nova.compute.manager [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Received event network-changed-ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 727.083843] env[61964]: DEBUG nova.compute.manager [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Refreshing instance network info cache due to event network-changed-ef8e5acc-8a42-44d0-84c4-57803733186b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 727.083974] env[61964]: DEBUG oslo_concurrency.lockutils [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] Acquiring lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.084729] env[61964]: DEBUG oslo_concurrency.lockutils [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] Acquired lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.084973] env[61964]: DEBUG nova.network.neutron [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Refreshing network info cache for port ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.201019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.201019] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 727.204206] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.224s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.209025] env[61964]: DEBUG nova.objects.instance [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lazy-loading 'resources' on Instance uuid d99859dc-2206-42d9-ae6d-8294fbd6942c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.334839] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a6038e-ccef-9d96-2f17-4ddcfe19b368, 'name': SearchDatastore_Task, 'duration_secs': 0.030225} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.335174] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.335426] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.335677] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.335935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.336021] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.336275] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63c60dd5-7a5c-4d0d-b314-618f1fc742d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.347298] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.347617] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 727.348436] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2513818-bb0c-449e-a6ce-9ab7e17ebf14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.354444] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 727.354444] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523d8502-b514-0b86-276c-55b04f91db1f" [ 727.354444] env[61964]: _type = "Task" [ 727.354444] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.363178] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523d8502-b514-0b86-276c-55b04f91db1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.372014] env[61964]: DEBUG nova.network.neutron [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.397922] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040749, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.400143] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 727.408230] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Successfully updated port: c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 727.614930] env[61964]: DEBUG nova.compute.manager [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Received event network-changed-8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 727.614930] env[61964]: DEBUG nova.compute.manager [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Refreshing instance network info cache due to event network-changed-8b2c44a6-df10-43fa-9afa-85167edbb650. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 727.615345] env[61964]: DEBUG oslo_concurrency.lockutils [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.615345] env[61964]: DEBUG oslo_concurrency.lockutils [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.615451] env[61964]: DEBUG nova.network.neutron [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Refreshing network info cache for port 8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.707857] env[61964]: DEBUG nova.compute.utils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.709238] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 727.709406] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.833451] env[61964]: DEBUG nova.policy [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'deb4b4579def4336bc9346cbfbc2dbc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a81b6d03c206476bbc0520d5880c19dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.868280] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523d8502-b514-0b86-276c-55b04f91db1f, 'name': SearchDatastore_Task, 'duration_secs': 0.0472} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.873744] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-249e55c9-d20a-4b2f-b1f7-bd113cfdcd42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.880151] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.894611] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 727.894611] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52731ae3-95a0-b647-c17a-1abdbd47ecc5" [ 727.894611] env[61964]: _type = "Task" [ 727.894611] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.901613] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.870016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.904715] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0ea19f52-b700-405f-8c3c-fbaa3e17b36a/0ea19f52-b700-405f-8c3c-fbaa3e17b36a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 727.908283] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 727.916605] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e62b38b8-1ce3-4105-844a-a5e796c7dd4f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.918696] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52731ae3-95a0-b647-c17a-1abdbd47ecc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.921659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.921844] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquired lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.921918] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.926658] env[61964]: DEBUG nova.network.neutron [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updated VIF entry in instance network info cache for port ef8e5acc-8a42-44d0-84c4-57803733186b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 727.926996] env[61964]: DEBUG nova.network.neutron [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating instance_info_cache with network_info: [{"id": "ef8e5acc-8a42-44d0-84c4-57803733186b", "address": "fa:16:3e:2e:93:49", "network": {"id": "c3e10648-bb91-4e2c-9f64-9b0479002071", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-192055358-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a99659516ad4274b631674f20b36280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e5acc-8a", "ovs_interfaceid": "ef8e5acc-8a42-44d0-84c4-57803733186b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.933700] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 727.933700] env[61964]: value = "task-1040750" [ 727.933700] env[61964]: _type = "Task" [ 727.933700] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.938599] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.949503] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040750, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.188997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cf025d-711d-4d07-bf89-9712318d7fa5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.199846] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18c9b17-e41c-449b-a2cc-cba270671c2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.230645] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 728.233986] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Successfully created port: b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.236454] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707273b2-b378-4e0f-9a1c-ba82e64a1aef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.245226] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1534f621-2674-4556-b1ff-ce2703756cda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.261820] env[61964]: DEBUG nova.compute.provider_tree [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.386021] env[61964]: DEBUG nova.network.neutron [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updated VIF entry in instance network info cache for port 8b2c44a6-df10-43fa-9afa-85167edbb650. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 728.386021] env[61964]: DEBUG nova.network.neutron [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.413640] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52731ae3-95a0-b647-c17a-1abdbd47ecc5, 'name': SearchDatastore_Task, 'duration_secs': 0.018069} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.414891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.415222] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 728.416282] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119dd6e5-fcee-4d66-96ed-f2b614a67a58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.419137] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c81f83a-31ff-4173-b55f-83890ba525d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.441513] env[61964]: DEBUG oslo_concurrency.lockutils [req-7700b46c-4990-4963-a2b0-c6b682653bd9 req-534f0164-859f-4386-a387-737309c5147c service nova] Releasing lock "refresh_cache-6ee845c3-5f79-4704-8b7b-cd3770202647" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.447483] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ec86c2-3b5f-4f95-bd29-e9105c24d37c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.451207] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 728.451207] env[61964]: value = "task-1040751" [ 728.451207] env[61964]: _type = "Task" [ 728.451207] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.465114] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040750, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191546} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.465563] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 728.472008] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 728.473292] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228ebf5e-d941-4a92-ad19-baae2a369493 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.481536] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.501553] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] 0ea19f52-b700-405f-8c3c-fbaa3e17b36a/0ea19f52-b700-405f-8c3c-fbaa3e17b36a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.502585] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.504918] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8cd6d7a-2d81-4389-a483-245dce2a68cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.529524] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 728.529524] env[61964]: value = "task-1040752" [ 728.529524] env[61964]: _type = "Task" [ 728.529524] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.538947] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040752, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.693900] env[61964]: DEBUG nova.network.neutron [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Updating instance_info_cache with network_info: [{"id": "c069b04e-3855-455f-8f6e-3efc1500a988", "address": "fa:16:3e:a1:e1:4f", "network": {"id": "ef6325ff-a85d-4efe-995f-b29de0991d2f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2029869707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da34d9d49bcc4fab81bbeb4cdf4e3992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069b04e-38", "ovs_interfaceid": "c069b04e-3855-455f-8f6e-3efc1500a988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.709048] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.770048] env[61964]: DEBUG nova.scheduler.client.report [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.889845] env[61964]: DEBUG oslo_concurrency.lockutils [req-c10ddb1f-1134-468f-b5c3-1353e49cf2cd req-69386772-ea0c-4ec7-81ff-c68dc0afe589 service nova] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.961690] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.974916] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 728.976031] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9ace82c-7914-4a2e-ac0c-9373d68d4251 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.982921] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 728.982921] env[61964]: value = "task-1040753" [ 728.982921] env[61964]: _type = "Task" [ 728.982921] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.992955] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.040204] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040752, 'name': ReconfigVM_Task, 'duration_secs': 0.285493} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.040600] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Reconfigured VM instance instance-0000002e to attach disk [datastore1] 0ea19f52-b700-405f-8c3c-fbaa3e17b36a/0ea19f52-b700-405f-8c3c-fbaa3e17b36a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.041333] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7e7433e-fd9f-4bc0-9548-a7dd2dd50456 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.049313] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 729.049313] env[61964]: value = "task-1040754" [ 729.049313] env[61964]: _type = "Task" [ 729.049313] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.058505] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040754, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.153265] env[61964]: DEBUG nova.compute.manager [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Received event network-vif-plugged-c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 729.153500] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Acquiring lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.153720] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.153941] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.154197] env[61964]: DEBUG nova.compute.manager [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] No waiting events found dispatching network-vif-plugged-c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 729.154313] env[61964]: WARNING nova.compute.manager [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Received unexpected event network-vif-plugged-c069b04e-3855-455f-8f6e-3efc1500a988 for instance with vm_state building and task_state spawning. [ 729.154553] env[61964]: DEBUG nova.compute.manager [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Received event network-changed-c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 729.154760] env[61964]: DEBUG nova.compute.manager [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Refreshing instance network info cache due to event network-changed-c069b04e-3855-455f-8f6e-3efc1500a988. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 729.154939] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Acquiring lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.196957] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Releasing lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.197352] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Instance network_info: |[{"id": "c069b04e-3855-455f-8f6e-3efc1500a988", "address": "fa:16:3e:a1:e1:4f", "network": {"id": "ef6325ff-a85d-4efe-995f-b29de0991d2f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2029869707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da34d9d49bcc4fab81bbeb4cdf4e3992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069b04e-38", "ovs_interfaceid": "c069b04e-3855-455f-8f6e-3efc1500a988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 729.197702] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Acquired lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.197925] env[61964]: DEBUG nova.network.neutron [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Refreshing network info cache for port c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.199644] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:e1:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57c65f87-60fd-4882-ab30-31db49131b46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c069b04e-3855-455f-8f6e-3efc1500a988', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.208612] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Creating folder: Project (da34d9d49bcc4fab81bbeb4cdf4e3992). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.212059] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5111c963-b1e7-4da2-ae2d-0a0173594c6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.213903] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.214245] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 729.214620] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.215620] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.215846] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.216061] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.216257] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.216447] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.216631] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 729.217034] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 729.228178] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Created folder: Project (da34d9d49bcc4fab81bbeb4cdf4e3992) in parent group-v230360. [ 729.228178] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Creating folder: Instances. Parent ref: group-v230430. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 729.228178] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6723737c-153c-40ee-b514-569bf0358ec9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.241944] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Created folder: Instances in parent group-v230430. [ 729.246058] env[61964]: DEBUG oslo.service.loopingcall [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.246058] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 729.246058] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5b20d2b-0969-4900-a25f-54d5308f88ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.261490] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 729.271284] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.271284] env[61964]: value = "task-1040757" [ 729.271284] env[61964]: _type = "Task" [ 729.271284] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.276209] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.072s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.278768] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.207s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.279382] env[61964]: DEBUG nova.objects.instance [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lazy-loading 'resources' on Instance uuid 6ced6800-db29-4766-8d83-b63b50d5fcc5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 729.289100] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040757, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.294961] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 729.295204] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.295367] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 729.295591] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.296514] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 729.296514] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 729.296514] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 729.296514] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 729.296514] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 729.296770] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 729.296770] env[61964]: DEBUG nova.virt.hardware [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 729.298929] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd937ef-02e0-4868-8238-f7424a238adc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.305457] env[61964]: INFO nova.scheduler.client.report [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted allocations for instance d99859dc-2206-42d9-ae6d-8294fbd6942c [ 729.312695] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0c32d6-850e-4be6-ba4b-6d5afbf6615a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.464761] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040751, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.496382] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040753, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.540683] env[61964]: DEBUG nova.network.neutron [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Updated VIF entry in instance network info cache for port c069b04e-3855-455f-8f6e-3efc1500a988. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 729.541209] env[61964]: DEBUG nova.network.neutron [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Updating instance_info_cache with network_info: [{"id": "c069b04e-3855-455f-8f6e-3efc1500a988", "address": "fa:16:3e:a1:e1:4f", "network": {"id": "ef6325ff-a85d-4efe-995f-b29de0991d2f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-2029869707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "da34d9d49bcc4fab81bbeb4cdf4e3992", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57c65f87-60fd-4882-ab30-31db49131b46", "external-id": "nsx-vlan-transportzone-610", "segmentation_id": 610, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc069b04e-38", "ovs_interfaceid": "c069b04e-3855-455f-8f6e-3efc1500a988", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.562238] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040754, 'name': Rename_Task, 'duration_secs': 0.151397} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.562582] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 729.562855] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51b2dde3-b91e-4967-892b-c08202de9119 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.575530] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 729.575530] env[61964]: value = "task-1040758" [ 729.575530] env[61964]: _type = "Task" [ 729.575530] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.587395] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.720706] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.785331] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040757, 'name': CreateVM_Task, 'duration_secs': 0.475551} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.786142] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 729.787267] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.787577] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.788028] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.788398] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a626d27-4cd7-43df-a6df-ef31cb28fa15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.793804] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 729.793804] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52903bea-d036-cad6-ec6d-040466028ff3" [ 729.793804] env[61964]: _type = "Task" [ 729.793804] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.803024] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52903bea-d036-cad6-ec6d-040466028ff3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.820809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-98ae84b3-a321-44d0-b8bb-22721c3d1d6b tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "d99859dc-2206-42d9-ae6d-8294fbd6942c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.349s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.896121] env[61964]: DEBUG nova.compute.manager [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Received event network-vif-plugged-b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 729.896356] env[61964]: DEBUG oslo_concurrency.lockutils [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] Acquiring lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.896553] env[61964]: DEBUG oslo_concurrency.lockutils [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.896720] env[61964]: DEBUG oslo_concurrency.lockutils [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.896918] env[61964]: DEBUG nova.compute.manager [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] No waiting events found dispatching network-vif-plugged-b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 729.897111] env[61964]: WARNING nova.compute.manager [req-81bc41d8-3b12-48a1-bee2-afdc7c517193 req-08b66f77-31f8-4173-a1e2-506582af6198 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Received unexpected event network-vif-plugged-b5773bca-cae3-4d80-9784-11117858828f for instance with vm_state building and task_state spawning. [ 729.964563] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040751, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.161554} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.967886] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 729.967886] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.969994] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e5db1d7-5254-4d3d-98d8-bae6997e50a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.980175] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 729.980175] env[61964]: value = "task-1040759" [ 729.980175] env[61964]: _type = "Task" [ 729.980175] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.990447] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.000011] env[61964]: DEBUG oslo_vmware.api [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040753, 'name': PowerOnVM_Task, 'duration_secs': 0.685787} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.000409] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 730.000662] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1f2710ff-f04b-485e-b4b0-d7797cd52e30 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance '65566c20-6d69-471c-b098-3c30c01d9955' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 730.044768] env[61964]: DEBUG oslo_concurrency.lockutils [req-9bd7aaa4-2dcc-41b1-a63d-e23d854c59d5 req-6f7675dc-9d5e-4beb-8c8e-96ce48f766ad service nova] Releasing lock "refresh_cache-e5e7b687-5691-4b1e-b113-2b8e83b23662" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.058908] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Successfully updated port: b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 730.088126] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040758, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.223743] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88115f66-80dd-4840-ae57-00e905f70c10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.233881] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e92aa0d-252c-4f2d-a69d-ccc801e7b84a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.266180] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2072ab-7ce1-4a21-9b99-977d6815acf3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.274310] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab38ddf2-b3ca-4801-9bfc-1dd89fd5aaa9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.994444] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.994753] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquired lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.994753] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.015559] env[61964]: DEBUG nova.compute.provider_tree [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.025466] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52903bea-d036-cad6-ec6d-040466028ff3, 'name': SearchDatastore_Task, 'duration_secs': 0.011134} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.025753] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121365} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.027170] env[61964]: DEBUG nova.scheduler.client.report [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.030279] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.030516] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.031819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.031819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.031819] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.031819] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.034736] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6cc39deb-e951-48bf-b273-c22d37bbc75b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.037066] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2872da52-be77-4235-a794-75c12324fc12 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.039961] env[61964]: DEBUG oslo_vmware.api [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040758, 'name': PowerOnVM_Task, 'duration_secs': 0.558076} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.040240] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 731.040439] env[61964]: INFO nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Took 8.33 seconds to spawn the instance on the hypervisor. [ 731.040624] env[61964]: DEBUG nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 731.042398] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76910095-def4-4e68-9602-9b01ce979f16 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.062140] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.064193] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea86103a-ed4c-4465-a06e-bf397d906cf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.079082] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.079303] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 731.081750] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8eb2cfe-aa22-44ea-b92d-7278e7fefa2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.091101] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 731.091101] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520fad5e-46d5-ff94-34a5-3f8ce63e57d6" [ 731.091101] env[61964]: _type = "Task" [ 731.091101] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.096070] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 731.096070] env[61964]: value = "task-1040760" [ 731.096070] env[61964]: _type = "Task" [ 731.096070] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.102718] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520fad5e-46d5-ff94-34a5-3f8ce63e57d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.110270] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040760, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.536514] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.258s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.542024] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.356s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.542024] env[61964]: INFO nova.compute.claims [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.553914] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.560403] env[61964]: INFO nova.scheduler.client.report [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted allocations for instance 6ced6800-db29-4766-8d83-b63b50d5fcc5 [ 731.604282] env[61964]: INFO nova.compute.manager [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Took 35.82 seconds to build instance. [ 731.617932] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040760, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.622330] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520fad5e-46d5-ff94-34a5-3f8ce63e57d6, 'name': SearchDatastore_Task, 'duration_secs': 0.011448} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.625882] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abe5d4cb-dac6-4250-b241-24f8a02aa4a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.633491] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 731.633491] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226f822-4f71-f2aa-645c-0fe92a56c841" [ 731.633491] env[61964]: _type = "Task" [ 731.633491] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.644334] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226f822-4f71-f2aa-645c-0fe92a56c841, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.755447] env[61964]: DEBUG nova.network.neutron [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Updating instance_info_cache with network_info: [{"id": "b5773bca-cae3-4d80-9784-11117858828f", "address": "fa:16:3e:3c:3a:06", "network": {"id": "4be1b93c-f7e2-434f-837e-22ce776f32c5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1528573541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a81b6d03c206476bbc0520d5880c19dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5773bca-ca", "ovs_interfaceid": "b5773bca-cae3-4d80-9784-11117858828f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.009942] env[61964]: DEBUG nova.compute.manager [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Received event network-changed-b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 732.010772] env[61964]: DEBUG nova.compute.manager [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Refreshing instance network info cache due to event network-changed-b5773bca-cae3-4d80-9784-11117858828f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 732.010772] env[61964]: DEBUG oslo_concurrency.lockutils [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] Acquiring lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.071914] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7073fe68-cc9e-4da9-85f8-1a39f7718777 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "6ced6800-db29-4766-8d83-b63b50d5fcc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.519s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.110901] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5bff0dc8-9301-4234-898f-0ce6a7613dc4 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.565s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.111238] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040760, 'name': ReconfigVM_Task, 'duration_secs': 0.600344} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.112614] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Reconfigured VM instance instance-0000002b to attach disk [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5/df62ed65-0a89-4f04-9b5a-f5f4214e6bd5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.113368] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-359a71ef-5902-4c30-b4a4-44829d3d0669 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.122635] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 732.122635] env[61964]: value = "task-1040761" [ 732.122635] env[61964]: _type = "Task" [ 732.122635] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.132424] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040761, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.143819] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226f822-4f71-f2aa-645c-0fe92a56c841, 'name': SearchDatastore_Task, 'duration_secs': 0.011392} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.144148] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.144420] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] e5e7b687-5691-4b1e-b113-2b8e83b23662/e5e7b687-5691-4b1e-b113-2b8e83b23662.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 732.144714] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b982b12-70ee-4baa-9b24-7e75564c499c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.152717] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 732.152717] env[61964]: value = "task-1040762" [ 732.152717] env[61964]: _type = "Task" [ 732.152717] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.163880] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040762, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.258444] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Releasing lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.258444] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Instance network_info: |[{"id": "b5773bca-cae3-4d80-9784-11117858828f", "address": "fa:16:3e:3c:3a:06", "network": {"id": "4be1b93c-f7e2-434f-837e-22ce776f32c5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1528573541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a81b6d03c206476bbc0520d5880c19dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5773bca-ca", "ovs_interfaceid": "b5773bca-cae3-4d80-9784-11117858828f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 732.258720] env[61964]: DEBUG oslo_concurrency.lockutils [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] Acquired lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.258909] env[61964]: DEBUG nova.network.neutron [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Refreshing network info cache for port b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.260211] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:3a:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5773bca-cae3-4d80-9784-11117858828f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.267755] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Creating folder: Project (a81b6d03c206476bbc0520d5880c19dc). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.268795] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e44ff60-6238-42f0-8ae9-ed275643f3d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.282727] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Created folder: Project (a81b6d03c206476bbc0520d5880c19dc) in parent group-v230360. [ 732.283013] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Creating folder: Instances. Parent ref: group-v230433. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.283305] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15135e2a-2288-4f30-9bb8-9b4e0313cab0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.292692] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Created folder: Instances in parent group-v230433. [ 732.292941] env[61964]: DEBUG oslo.service.loopingcall [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.293156] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 732.293365] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-619b4f2a-50de-4d43-bd1b-7a5e2de89eba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.312512] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.312512] env[61964]: value = "task-1040765" [ 732.312512] env[61964]: _type = "Task" [ 732.312512] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.322328] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040765, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.555791] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.556116] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.556341] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.556620] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.556764] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.559999] env[61964]: INFO nova.compute.manager [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Terminating instance [ 732.614960] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 732.643340] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040761, 'name': Rename_Task, 'duration_secs': 0.164767} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.643657] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.643915] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f93fd833-ec53-42a7-aa1f-7d11e59f2718 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.658355] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 732.658355] env[61964]: value = "task-1040766" [ 732.658355] env[61964]: _type = "Task" [ 732.658355] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.675821] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040762, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.684942] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040766, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.825355] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040765, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.056025] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61df24e6-dfb7-44da-ae7f-7b1fd3989ece {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.066905] env[61964]: DEBUG nova.compute.manager [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 733.067591] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.070117] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7213aa-fcdd-4a31-98af-909b71d7af95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.072758] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ffc315-5c74-47af-9587-ff3ed75d40b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.082127] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 733.113673] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8994d86f-b43f-424a-812a-7f8405749a5b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.117456] env[61964]: DEBUG nova.network.neutron [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Port 03f523b6-4fc2-4fe2-a751-d33c00f6849f binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 733.117456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.117456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.117456] env[61964]: DEBUG nova.network.neutron [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.121013] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e5b7d1-dab3-48b7-91ee-0e7c904d355a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.133356] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 733.133356] env[61964]: value = "task-1040767" [ 733.133356] env[61964]: _type = "Task" [ 733.133356] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.137561] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e0f1df-98be-4934-859b-2abe28a0f44e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.158386] env[61964]: DEBUG nova.compute.provider_tree [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.163997] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.168350] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.170266] env[61964]: DEBUG nova.scheduler.client.report [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.184497] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040762, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696085} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.189101] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] e5e7b687-5691-4b1e-b113-2b8e83b23662/e5e7b687-5691-4b1e-b113-2b8e83b23662.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 733.189101] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 733.189101] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040766, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.190984] env[61964]: DEBUG nova.network.neutron [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Updated VIF entry in instance network info cache for port b5773bca-cae3-4d80-9784-11117858828f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 733.190984] env[61964]: DEBUG nova.network.neutron [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Updating instance_info_cache with network_info: [{"id": "b5773bca-cae3-4d80-9784-11117858828f", "address": "fa:16:3e:3c:3a:06", "network": {"id": "4be1b93c-f7e2-434f-837e-22ce776f32c5", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1528573541-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a81b6d03c206476bbc0520d5880c19dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5773bca-ca", "ovs_interfaceid": "b5773bca-cae3-4d80-9784-11117858828f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.192034] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-863de119-f526-42fe-b20d-089dc2b1fa71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.204687] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 733.204687] env[61964]: value = "task-1040768" [ 733.204687] env[61964]: _type = "Task" [ 733.204687] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.216763] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.323730] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040765, 'name': CreateVM_Task, 'duration_secs': 0.751793} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.323904] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 733.324678] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.324878] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.325208] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 733.325487] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd93a280-8562-4049-9621-8d1a8d02063f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.332223] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 733.332223] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f60e88-3858-ab66-c4f6-43c1eaba017b" [ 733.332223] env[61964]: _type = "Task" [ 733.332223] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.341554] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f60e88-3858-ab66-c4f6-43c1eaba017b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.652437] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040767, 'name': PowerOffVM_Task, 'duration_secs': 0.24013} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.652872] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 733.653040] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 733.653354] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c8076ed-40c2-4afa-a43d-81e2b05f54e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.670630] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "062da016-f4d6-424d-90e6-5903940e2daa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.670867] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.675287] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.136s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.675797] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.678236] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.993s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.679734] env[61964]: INFO nova.compute.claims [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.687843] env[61964]: DEBUG oslo_vmware.api [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040766, 'name': PowerOnVM_Task, 'duration_secs': 0.686572} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.688103] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 733.688320] env[61964]: DEBUG nova.compute.manager [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 733.689175] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4979543-4f54-443b-9add-2bc6040f0e0a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.699296] env[61964]: DEBUG oslo_concurrency.lockutils [req-87be5c49-fc33-4408-a344-30fc555a6457 req-111d776c-ff22-4be8-a21d-fa7ee1c5dd47 service nova] Releasing lock "refresh_cache-b50e683f-79b2-419a-ac95-94908adb2b3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.704979] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.705222] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.715432] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082985} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.715738] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.716587] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd102dc-ce34-443f-a5ab-d38bc2121d64 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.720614] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 733.720897] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 733.721124] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore1] 0ea19f52-b700-405f-8c3c-fbaa3e17b36a {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 733.722618] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4397c4b3-4cf8-4bac-8a5b-4cc608260811 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.744864] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] e5e7b687-5691-4b1e-b113-2b8e83b23662/e5e7b687-5691-4b1e-b113-2b8e83b23662.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.748210] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ceb6c6f-cf49-486d-8936-f8fa4a556fac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.764324] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 733.764324] env[61964]: value = "task-1040770" [ 733.764324] env[61964]: _type = "Task" [ 733.764324] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.769876] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 733.769876] env[61964]: value = "task-1040771" [ 733.769876] env[61964]: _type = "Task" [ 733.769876] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.775972] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.782221] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040771, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.843423] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f60e88-3858-ab66-c4f6-43c1eaba017b, 'name': SearchDatastore_Task, 'duration_secs': 0.023732} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.843737] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.843976] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 733.844233] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.844423] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.844660] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 733.844892] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbc799d6-3e4a-4d22-9854-1298cc29c957 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.854799] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 733.854976] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 733.855711] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23330fcf-7e0d-4606-8e80-cff38952f6f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.862207] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 733.862207] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522fd822-822d-346e-32a5-a6e9f197ddea" [ 733.862207] env[61964]: _type = "Task" [ 733.862207] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.871946] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522fd822-822d-346e-32a5-a6e9f197ddea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.901399] env[61964]: DEBUG nova.network.neutron [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.185446] env[61964]: DEBUG nova.compute.utils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.190486] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 734.190670] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.210129] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.245392] env[61964]: DEBUG nova.policy [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '576d1881b73b43958ebbab05d8e523ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '540083b7cc4a4b8aa25714afc2add861', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.277705] env[61964]: DEBUG oslo_vmware.api [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421472} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.278402] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.278616] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 734.278802] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.278981] env[61964]: INFO nova.compute.manager [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Took 1.21 seconds to destroy the instance on the hypervisor. [ 734.279242] env[61964]: DEBUG oslo.service.loopingcall [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.279436] env[61964]: DEBUG nova.compute.manager [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 734.279525] env[61964]: DEBUG nova.network.neutron [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.284154] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040771, 'name': ReconfigVM_Task, 'duration_secs': 0.322965} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.284798] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Reconfigured VM instance instance-0000002f to attach disk [datastore1] e5e7b687-5691-4b1e-b113-2b8e83b23662/e5e7b687-5691-4b1e-b113-2b8e83b23662.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.285480] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89746631-0f3b-471f-84db-fae1d6216441 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.293011] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 734.293011] env[61964]: value = "task-1040772" [ 734.293011] env[61964]: _type = "Task" [ 734.293011] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.301311] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040772, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.377028] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522fd822-822d-346e-32a5-a6e9f197ddea, 'name': SearchDatastore_Task, 'duration_secs': 0.03518} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.377028] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16337a2a-87e5-4bc7-849b-1db856d7286a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.380664] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 734.380664] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d81677-17c0-0c6f-daec-d9d96171458b" [ 734.380664] env[61964]: _type = "Task" [ 734.380664] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.390907] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d81677-17c0-0c6f-daec-d9d96171458b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.403688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.692482] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.777024] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Successfully created port: 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.809360] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040772, 'name': Rename_Task, 'duration_secs': 0.174619} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.809972] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 734.810660] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c42eab7-d1ba-4071-b35f-ed9371e38121 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.818789] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 734.818789] env[61964]: value = "task-1040773" [ 734.818789] env[61964]: _type = "Task" [ 734.818789] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.827464] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.886670] env[61964]: DEBUG nova.compute.manager [req-ff88277a-93c8-4d1e-9485-aab70c5a8689 req-13f66aa2-617b-44d2-9e0d-b18f4dbeef2a service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Received event network-vif-deleted-33b1704a-2711-4fb6-bae6-94dc5716b4d3 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 734.886881] env[61964]: INFO nova.compute.manager [req-ff88277a-93c8-4d1e-9485-aab70c5a8689 req-13f66aa2-617b-44d2-9e0d-b18f4dbeef2a service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Neutron deleted interface 33b1704a-2711-4fb6-bae6-94dc5716b4d3; detaching it from the instance and deleting it from the info cache [ 734.887196] env[61964]: DEBUG nova.network.neutron [req-ff88277a-93c8-4d1e-9485-aab70c5a8689 req-13f66aa2-617b-44d2-9e0d-b18f4dbeef2a service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.901373] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d81677-17c0-0c6f-daec-d9d96171458b, 'name': SearchDatastore_Task, 'duration_secs': 0.011937} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.901646] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.901920] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] b50e683f-79b2-419a-ac95-94908adb2b3e/b50e683f-79b2-419a-ac95-94908adb2b3e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 734.902354] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9875bc3e-1b1f-4d05-85e2-f512f8c47185 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.907524] env[61964]: DEBUG nova.compute.manager [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61964) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 734.907740] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.912894] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 734.912894] env[61964]: value = "task-1040774" [ 734.912894] env[61964]: _type = "Task" [ 734.912894] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.928579] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.161805] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df666c6-ad6d-4b4f-a8c6-b6b0f1445ead {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.172650] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c96ac98-be20-4c5b-863e-d09325dcd690 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.211927] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d0e8cf-c298-47f7-9ca8-8d3ef5d3ab59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.221935] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b608ad-68dd-4570-989a-349b77c7c3c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.239090] env[61964]: DEBUG nova.compute.provider_tree [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.255723] env[61964]: DEBUG nova.network.neutron [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.331186] env[61964]: DEBUG oslo_vmware.api [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040773, 'name': PowerOnVM_Task, 'duration_secs': 0.499215} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.331603] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 735.331869] env[61964]: INFO nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Took 8.78 seconds to spawn the instance on the hypervisor. [ 735.332216] env[61964]: DEBUG nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 735.333517] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbdafe5-6566-4ee3-bfcf-cc2e2fc3ef86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.397617] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62eaf2f0-f1d5-4ccd-8504-185fc1edb9ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.413498] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfe0c47-f005-4180-b90b-49183ae0a1a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.428182] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.428401] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.428659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.429595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.429595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.432029] env[61964]: INFO nova.compute.manager [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Terminating instance [ 735.438960] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040774, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.459801] env[61964]: DEBUG nova.compute.manager [req-ff88277a-93c8-4d1e-9485-aab70c5a8689 req-13f66aa2-617b-44d2-9e0d-b18f4dbeef2a service nova] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Detach interface failed, port_id=33b1704a-2711-4fb6-bae6-94dc5716b4d3, reason: Instance 0ea19f52-b700-405f-8c3c-fbaa3e17b36a could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 735.716183] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.742241] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 735.742514] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.742685] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.742904] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.743072] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.743226] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 735.743428] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 735.743590] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 735.743757] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 735.743924] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 735.744112] env[61964]: DEBUG nova.virt.hardware [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 735.745013] env[61964]: DEBUG nova.scheduler.client.report [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.748639] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb6adad-7681-438f-854d-7fbfa6b5cca1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.757519] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a7d3b3-b55f-467d-a3c1-a454f952ff83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.761727] env[61964]: INFO nova.compute.manager [-] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Took 1.48 seconds to deallocate network for instance. [ 735.859181] env[61964]: INFO nova.compute.manager [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Took 33.65 seconds to build instance. [ 735.918956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.919211] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.935047] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58425} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.935166] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] b50e683f-79b2-419a-ac95-94908adb2b3e/b50e683f-79b2-419a-ac95-94908adb2b3e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 735.935361] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 735.935773] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e877b059-d469-4c84-8b0b-80f04fdd548c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.939928] env[61964]: DEBUG nova.compute.manager [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 735.940165] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 735.940960] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fe5273-e850-451f-a8c3-fe80ee4dd858 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.944891] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 735.944891] env[61964]: value = "task-1040775" [ 735.944891] env[61964]: _type = "Task" [ 735.944891] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.950871] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 735.953672] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0baa94b-6844-4e5f-bf06-23f1e446ff72 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.955285] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.962056] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 735.962056] env[61964]: value = "task-1040776" [ 735.962056] env[61964]: _type = "Task" [ 735.962056] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.971441] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.253201] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.253745] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 736.256345] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.087s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.257865] env[61964]: INFO nova.compute.claims [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.275099] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.361285] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c40bf8ba-5996-492d-9af0-406b34e19495 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.281s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.458177] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117907} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.460016] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 736.460980] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bc87ba-e627-45c6-8bef-954ca7971b33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.485568] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040776, 'name': PowerOffVM_Task, 'duration_secs': 0.219069} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.497200] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] b50e683f-79b2-419a-ac95-94908adb2b3e/b50e683f-79b2-419a-ac95-94908adb2b3e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.497200] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 736.497200] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 736.498085] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ac068ec-9244-47a3-8e67-db3846299762 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.515199] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca6fddbf-a182-4536-aa79-ae50ee3b9685 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.528631] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 736.528631] env[61964]: value = "task-1040778" [ 736.528631] env[61964]: _type = "Task" [ 736.528631] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.532278] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Successfully updated port: 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.540057] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040778, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.586548] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 736.587189] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 736.587461] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] df62ed65-0a89-4f04-9b5a-f5f4214e6bd5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 736.588051] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-663ecd60-a7ce-4c20-8943-de17d93268b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.596023] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 736.596023] env[61964]: value = "task-1040779" [ 736.596023] env[61964]: _type = "Task" [ 736.596023] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.605118] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.762551] env[61964]: DEBUG nova.compute.utils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.766487] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 736.766487] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.837743] env[61964]: DEBUG nova.policy [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f99415b64be4c90ba2d7e8a7c01674e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c4bc7eec793453e8acff852da039375', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.863870] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 736.917934] env[61964]: DEBUG nova.compute.manager [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-vif-plugged-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 736.918197] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Acquiring lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.918444] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.918628] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.918806] env[61964]: DEBUG nova.compute.manager [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] No waiting events found dispatching network-vif-plugged-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 736.918980] env[61964]: WARNING nova.compute.manager [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received unexpected event network-vif-plugged-89204ea4-fd9b-4979-8aaf-f224046531fa for instance with vm_state building and task_state spawning. [ 736.919607] env[61964]: DEBUG nova.compute.manager [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 736.919800] env[61964]: DEBUG nova.compute.manager [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing instance network info cache due to event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 736.919998] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.920191] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.920363] env[61964]: DEBUG nova.network.neutron [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.939286] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.939421] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.939636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.939829] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.940012] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.942067] env[61964]: INFO nova.compute.manager [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Terminating instance [ 737.036613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.037131] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040778, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.106981] env[61964]: DEBUG oslo_vmware.api [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.368452} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.107303] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 737.107501] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 737.107700] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 737.107910] env[61964]: INFO nova.compute.manager [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 737.108209] env[61964]: DEBUG oslo.service.loopingcall [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.108419] env[61964]: DEBUG nova.compute.manager [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 737.108514] env[61964]: DEBUG nova.network.neutron [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 737.223528] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Successfully created port: b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.269483] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 737.394024] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.450033] env[61964]: DEBUG nova.compute.manager [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 737.450308] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 737.451701] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2525fa5-e219-4777-a448-e8163195a970 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.469252] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 737.469584] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1bd51ed-a3b9-4c94-9c40-9163534dced8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.482546] env[61964]: DEBUG nova.network.neutron [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.487561] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 737.487561] env[61964]: value = "task-1040780" [ 737.487561] env[61964]: _type = "Task" [ 737.487561] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.497543] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.538840] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040778, 'name': ReconfigVM_Task, 'duration_secs': 0.536605} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.538840] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Reconfigured VM instance instance-00000030 to attach disk [datastore1] b50e683f-79b2-419a-ac95-94908adb2b3e/b50e683f-79b2-419a-ac95-94908adb2b3e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.539339] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c82c7c7c-5d7c-4ffc-843d-2464a6373e93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.550385] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 737.550385] env[61964]: value = "task-1040781" [ 737.550385] env[61964]: _type = "Task" [ 737.550385] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.560177] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040781, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.651382] env[61964]: DEBUG nova.network.neutron [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.773523] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae9e861-a10e-4999-9cfe-5748e57d1b19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.784328] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80b0c66-e16d-49f0-9f20-b7a5cd9b7e1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.815425] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f686b13-2347-4493-8440-19b3bd34921f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.824583] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e900ebb-7a93-4af7-a64e-31012f475e63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.843019] env[61964]: DEBUG nova.compute.provider_tree [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.997760] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040780, 'name': PowerOffVM_Task, 'duration_secs': 0.198355} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.998056] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 737.998238] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 737.998497] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b10201b-d8c2-44c9-b9c1-a0b2231a3ad2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.065486] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040781, 'name': Rename_Task, 'duration_secs': 0.15421} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.067200] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 738.067658] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 738.067893] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 738.068107] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Deleting the datastore file [datastore1] e5e7b687-5691-4b1e-b113-2b8e83b23662 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 738.068396] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee3fe3f0-099b-4a15-8230-d3e02a410b96 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.070206] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44d5beb3-87bc-4cd9-814b-46ce7b782b33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.077934] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for the task: (returnval){ [ 738.077934] env[61964]: value = "task-1040784" [ 738.077934] env[61964]: _type = "Task" [ 738.077934] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.079597] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 738.079597] env[61964]: value = "task-1040783" [ 738.079597] env[61964]: _type = "Task" [ 738.079597] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.093392] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040783, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.093392] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.113051] env[61964]: DEBUG nova.network.neutron [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.153928] env[61964]: DEBUG oslo_concurrency.lockutils [req-5e4ffc5f-9b5e-4a0e-b037-6cb8a1367cc8 req-292f227a-a690-4b00-879d-110880d9f9c4 service nova] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.154088] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.154262] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.284007] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 738.309717] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 738.309974] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.310224] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 738.310460] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.310616] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 738.310770] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 738.310980] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 738.311220] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 738.311428] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 738.311599] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 738.311777] env[61964]: DEBUG nova.virt.hardware [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 738.312729] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77b8a62-df26-46f1-9a8f-147efffd64c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.321288] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863162f5-dd2f-4136-b063-387354057a45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.343194] env[61964]: DEBUG nova.scheduler.client.report [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 738.593163] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040783, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.596345] env[61964]: DEBUG oslo_vmware.api [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Task: {'id': task-1040784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259871} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.596670] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.596938] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 738.597177] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 738.597370] env[61964]: INFO nova.compute.manager [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Took 1.15 seconds to destroy the instance on the hypervisor. [ 738.597651] env[61964]: DEBUG oslo.service.loopingcall [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.597855] env[61964]: DEBUG nova.compute.manager [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 738.597952] env[61964]: DEBUG nova.network.neutron [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.616263] env[61964]: INFO nova.compute.manager [-] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Took 1.51 seconds to deallocate network for instance. [ 738.686564] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.854030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.854030] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 738.857781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.928s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.859592] env[61964]: DEBUG nova.objects.instance [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lazy-loading 'resources' on Instance uuid a2a25787-1426-4c09-892d-395e1becccc9 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.970497] env[61964]: DEBUG nova.network.neutron [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.007309] env[61964]: DEBUG nova.compute.manager [req-4710a830-16de-4ee4-83df-2491e877a7be req-2211db1b-c9fa-44f2-8661-74e345a86f3c service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Received event network-vif-deleted-c069b04e-3855-455f-8f6e-3efc1500a988 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 739.007309] env[61964]: INFO nova.compute.manager [req-4710a830-16de-4ee4-83df-2491e877a7be req-2211db1b-c9fa-44f2-8661-74e345a86f3c service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Neutron deleted interface c069b04e-3855-455f-8f6e-3efc1500a988; detaching it from the instance and deleting it from the info cache [ 739.007309] env[61964]: DEBUG nova.network.neutron [req-4710a830-16de-4ee4-83df-2491e877a7be req-2211db1b-c9fa-44f2-8661-74e345a86f3c service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.084501] env[61964]: DEBUG nova.compute.manager [req-d01c9e97-27c8-4d7c-b985-e2aa73fb2234 req-e3f081c9-725e-43f2-be8c-ea157440597a service nova] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Received event network-vif-deleted-6d1d176f-6488-4095-b0a2-52d2f6d99245 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 739.095725] env[61964]: DEBUG oslo_vmware.api [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040783, 'name': PowerOnVM_Task, 'duration_secs': 0.874473} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.096065] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 739.096291] env[61964]: INFO nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Took 9.83 seconds to spawn the instance on the hypervisor. [ 739.096481] env[61964]: DEBUG nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 739.097272] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5ffb6c-8ccc-4990-bfe7-156d1aaadc31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.123343] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.209224] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Successfully updated port: b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 739.359489] env[61964]: DEBUG nova.compute.utils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.360941] env[61964]: DEBUG nova.network.neutron [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.362154] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 739.362233] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.407271] env[61964]: DEBUG nova.policy [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd91cc01a405946a19f6d845dae44713b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab93cadadb54acbb22a742f37613af3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.479093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.479418] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance network_info: |[{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 739.481869] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:d3:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89204ea4-fd9b-4979-8aaf-f224046531fa', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.490497] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Creating folder: Project (540083b7cc4a4b8aa25714afc2add861). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.493147] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6712409e-c82f-434f-865e-4dedb85480a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.507895] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Created folder: Project (540083b7cc4a4b8aa25714afc2add861) in parent group-v230360. [ 739.508128] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Creating folder: Instances. Parent ref: group-v230436. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 739.508548] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fffcade-98c0-448b-a743-35cc06177846 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.510828] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-517a98a0-4b38-47e1-956d-8430fba78c20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.522323] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cca5403-a069-4e1b-ae27-6943bf0f813a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.534982] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Created folder: Instances in parent group-v230436. [ 739.535254] env[61964]: DEBUG oslo.service.loopingcall [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.535893] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 739.536165] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5914ae23-c0fb-43f8-adc4-f3f60cb77b8b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.564057] env[61964]: DEBUG nova.compute.manager [req-4710a830-16de-4ee4-83df-2491e877a7be req-2211db1b-c9fa-44f2-8661-74e345a86f3c service nova] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Detach interface failed, port_id=c069b04e-3855-455f-8f6e-3efc1500a988, reason: Instance e5e7b687-5691-4b1e-b113-2b8e83b23662 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 739.570637] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.570637] env[61964]: value = "task-1040787" [ 739.570637] env[61964]: _type = "Task" [ 739.570637] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.578445] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040787, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.620407] env[61964]: INFO nova.compute.manager [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Took 36.30 seconds to build instance. [ 739.715096] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.715259] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquired lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.715413] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.772332] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Successfully created port: ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.830403] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fe0935-de84-4f01-9e78-cf90a1e58966 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.838593] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef02119-5810-435b-8ace-6ec233fe478d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.869422] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 739.872607] env[61964]: INFO nova.compute.manager [-] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Took 1.27 seconds to deallocate network for instance. [ 739.875512] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f366b6b2-2d59-478c-b3a8-512f09a0a815 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.888431] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9b65a8-6a5a-4a9a-a5ae-2ed53e67e8c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.902364] env[61964]: DEBUG nova.compute.provider_tree [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.081297] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040787, 'name': CreateVM_Task, 'duration_secs': 0.356119} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.081495] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 740.082282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.082464] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.082894] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 740.083181] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-909d497d-5202-498f-856c-83def2593a56 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.088672] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 740.088672] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b8319-8d35-6f22-8cf5-05054b7c11b0" [ 740.088672] env[61964]: _type = "Task" [ 740.088672] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.101026] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b8319-8d35-6f22-8cf5-05054b7c11b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.123472] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7a91ad30-81a9-4938-80d9-bc8f9830c307 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.382s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.282131] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.382894] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.405490] env[61964]: DEBUG nova.scheduler.client.report [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 740.491675] env[61964]: DEBUG nova.network.neutron [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Updating instance_info_cache with network_info: [{"id": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "address": "fa:16:3e:76:ee:7a", "network": {"id": "be83488e-7327-41d0-86eb-5d57d91c3d94", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-900978750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c4bc7eec793453e8acff852da039375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb94b591f-98", "ovs_interfaceid": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.599901] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b8319-8d35-6f22-8cf5-05054b7c11b0, 'name': SearchDatastore_Task, 'duration_secs': 0.011701} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.600312] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.600549] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.600785] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.600938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.601130] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.601387] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40250ca3-9999-4983-83d2-e478c2e9872e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.611876] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.611876] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 740.612309] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49495627-e77c-4a08-bb0a-4b6bf4f2cd59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.620356] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 740.620356] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aba0d5-b30c-4da8-4e4a-5eb41912414f" [ 740.620356] env[61964]: _type = "Task" [ 740.620356] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.628435] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 740.631242] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aba0d5-b30c-4da8-4e4a-5eb41912414f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.880473] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 740.906844] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 740.907143] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.907314] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 740.907500] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.907652] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 740.907803] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 740.908014] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 740.908194] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 740.908373] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 740.908536] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 740.908711] env[61964]: DEBUG nova.virt.hardware [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 740.909601] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c61df1-c074-489d-bd0c-decbb977e8ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.913037] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.915292] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.773s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.916863] env[61964]: INFO nova.compute.claims [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.922887] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bb22dd-ea45-4af0-ab9a-d13a4cf32425 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.939370] env[61964]: INFO nova.scheduler.client.report [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Deleted allocations for instance a2a25787-1426-4c09-892d-395e1becccc9 [ 740.993978] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Releasing lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.994795] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Instance network_info: |[{"id": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "address": "fa:16:3e:76:ee:7a", "network": {"id": "be83488e-7327-41d0-86eb-5d57d91c3d94", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-900978750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c4bc7eec793453e8acff852da039375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb94b591f-98", "ovs_interfaceid": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 740.994999] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:ee:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '419a5b3f-4c6f-4168-9def-746b4d8c5c24', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b94b591f-9808-4308-b7b6-739f0b48f4fa', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.002716] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Creating folder: Project (9c4bc7eec793453e8acff852da039375). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 741.003370] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa22eddb-dede-4f60-bbd4-ef6edaea2b97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.016252] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Created folder: Project (9c4bc7eec793453e8acff852da039375) in parent group-v230360. [ 741.016486] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Creating folder: Instances. Parent ref: group-v230439. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 741.016753] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f565e74e-08ac-44cc-88dc-6564079ef8c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.028808] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Created folder: Instances in parent group-v230439. [ 741.029260] env[61964]: DEBUG oslo.service.loopingcall [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.029550] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 741.029809] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83d28377-33b2-42de-af5d-bd3f4608a859 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.052280] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 741.052280] env[61964]: value = "task-1040790" [ 741.052280] env[61964]: _type = "Task" [ 741.052280] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.063640] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040790, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.113250] env[61964]: DEBUG nova.compute.manager [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Received event network-vif-plugged-b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 741.113250] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Acquiring lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.113756] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.113756] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.113756] env[61964]: DEBUG nova.compute.manager [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] No waiting events found dispatching network-vif-plugged-b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 741.113893] env[61964]: WARNING nova.compute.manager [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Received unexpected event network-vif-plugged-b94b591f-9808-4308-b7b6-739f0b48f4fa for instance with vm_state building and task_state spawning. [ 741.113998] env[61964]: DEBUG nova.compute.manager [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Received event network-changed-b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 741.114134] env[61964]: DEBUG nova.compute.manager [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Refreshing instance network info cache due to event network-changed-b94b591f-9808-4308-b7b6-739f0b48f4fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 741.114345] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Acquiring lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.114486] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Acquired lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.114648] env[61964]: DEBUG nova.network.neutron [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Refreshing network info cache for port b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 741.131056] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aba0d5-b30c-4da8-4e4a-5eb41912414f, 'name': SearchDatastore_Task, 'duration_secs': 0.009622} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.133984] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ddd062-e420-4da0-a8ea-69f9d3cb4ad6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.143834] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 741.143834] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52654073-e263-f64f-f34e-a38115d379b2" [ 741.143834] env[61964]: _type = "Task" [ 741.143834] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.149800] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.153349] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52654073-e263-f64f-f34e-a38115d379b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.193313] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "b50e683f-79b2-419a-ac95-94908adb2b3e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.193580] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.193797] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.193986] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.194362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.197087] env[61964]: INFO nova.compute.manager [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Terminating instance [ 741.446838] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cef7f4e2-cba8-4e1d-8a07-cdb4b04cbd06 tempest-AttachInterfacesV270Test-1031642760 tempest-AttachInterfacesV270Test-1031642760-project-member] Lock "a2a25787-1426-4c09-892d-395e1becccc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.541s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.567387] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040790, 'name': CreateVM_Task, 'duration_secs': 0.386952} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.567559] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 741.568298] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.568632] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.568847] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.569088] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c391a479-55c6-4ec7-98f8-c34c0f9bd0ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.575535] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 741.575535] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e2b8d-d720-fe68-5f88-60b5babc3849" [ 741.575535] env[61964]: _type = "Task" [ 741.575535] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.584785] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e2b8d-d720-fe68-5f88-60b5babc3849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.656783] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52654073-e263-f64f-f34e-a38115d379b2, 'name': SearchDatastore_Task, 'duration_secs': 0.010382} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.657088] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.657359] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 46435c68-f85a-4360-b2b5-6296afc33c3e/46435c68-f85a-4360-b2b5-6296afc33c3e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 741.657640] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87bb0e91-0b7a-411c-963c-11692788e74c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.667207] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 741.667207] env[61964]: value = "task-1040791" [ 741.667207] env[61964]: _type = "Task" [ 741.667207] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.678485] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.700619] env[61964]: DEBUG nova.compute.manager [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 741.700957] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 741.702030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cdf7aa-bf05-4d23-8543-9fdefadd9925 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.710815] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 741.711101] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3813e5a7-1510-48b1-9978-853656f617be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.721040] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 741.721040] env[61964]: value = "task-1040792" [ 741.721040] env[61964]: _type = "Task" [ 741.721040] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.730412] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.818910] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Successfully updated port: ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 742.011395] env[61964]: DEBUG nova.network.neutron [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Updated VIF entry in instance network info cache for port b94b591f-9808-4308-b7b6-739f0b48f4fa. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 742.011656] env[61964]: DEBUG nova.network.neutron [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Updating instance_info_cache with network_info: [{"id": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "address": "fa:16:3e:76:ee:7a", "network": {"id": "be83488e-7327-41d0-86eb-5d57d91c3d94", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-900978750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9c4bc7eec793453e8acff852da039375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "419a5b3f-4c6f-4168-9def-746b4d8c5c24", "external-id": "nsx-vlan-transportzone-656", "segmentation_id": 656, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb94b591f-98", "ovs_interfaceid": "b94b591f-9808-4308-b7b6-739f0b48f4fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.101465] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e2b8d-d720-fe68-5f88-60b5babc3849, 'name': SearchDatastore_Task, 'duration_secs': 0.010093} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.105657] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.106485] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.106889] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.107194] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.107515] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.108187] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8171888-c34b-47f2-81c7-0bbcdf4f474f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.125883] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.125883] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 742.125883] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cc5df7d-8654-46fa-a17c-c721d713d8a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.133517] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 742.133517] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41f89-5a1f-07b8-ee16-82984fdc71e7" [ 742.133517] env[61964]: _type = "Task" [ 742.133517] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.145754] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41f89-5a1f-07b8-ee16-82984fdc71e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.177094] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040791, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.232562] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040792, 'name': PowerOffVM_Task, 'duration_secs': 0.289445} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.232844] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 742.233043] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 742.233331] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff4a3de9-ff98-4162-9f18-1d9380977233 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.318089] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 742.318368] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 742.318753] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Deleting the datastore file [datastore1] b50e683f-79b2-419a-ac95-94908adb2b3e {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.319975] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-086e3246-0f85-44e9-852a-334a9ccf3e90 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.326528] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.326528] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquired lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.326528] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.329043] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for the task: (returnval){ [ 742.329043] env[61964]: value = "task-1040794" [ 742.329043] env[61964]: _type = "Task" [ 742.329043] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.345823] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040794, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.438203] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ed9c9f-4920-41cb-b977-2efa746ad2b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.446022] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c63c59-a80e-4ef2-8461-467d4f282fb6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.483161] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc46fa2-40de-43eb-8ac0-1e2cc7392beb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.492029] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c64f87-0ca6-4603-982a-bdb89e96a25e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.509968] env[61964]: DEBUG nova.compute.provider_tree [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 742.516691] env[61964]: DEBUG oslo_concurrency.lockutils [req-ecbb1a5b-bff3-45e3-a130-d344d32cbe54 req-037c415a-7062-4998-b4c1-1a9b99ce2532 service nova] Releasing lock "refresh_cache-85cd73df-a2f2-4f54-b984-832f9589a236" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.648024] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41f89-5a1f-07b8-ee16-82984fdc71e7, 'name': SearchDatastore_Task, 'duration_secs': 0.031977} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.648925] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d0cadcc-9ea4-4098-aaab-784d8b819002 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.655952] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 742.655952] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280f416-6d7c-9bd1-31b7-e632477cc6bb" [ 742.655952] env[61964]: _type = "Task" [ 742.655952] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.665142] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280f416-6d7c-9bd1-31b7-e632477cc6bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.677373] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040791, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513979} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.677731] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 46435c68-f85a-4360-b2b5-6296afc33c3e/46435c68-f85a-4360-b2b5-6296afc33c3e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 742.677987] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 742.678299] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da272de4-58b2-46e0-8851-edd1410436d1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.686289] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 742.686289] env[61964]: value = "task-1040795" [ 742.686289] env[61964]: _type = "Task" [ 742.686289] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.694773] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.844950] env[61964]: DEBUG oslo_vmware.api [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Task: {'id': task-1040794, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153166} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.845257] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.845452] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 742.845635] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.845816] env[61964]: INFO nova.compute.manager [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 742.846076] env[61964]: DEBUG oslo.service.loopingcall [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.846276] env[61964]: DEBUG nova.compute.manager [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 742.846368] env[61964]: DEBUG nova.network.neutron [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.909298] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.035062] env[61964]: ERROR nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [req-bdfc0937-ed0e-4cdf-9021-3c2e14febfb7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bdfc0937-ed0e-4cdf-9021-3c2e14febfb7"}]} [ 743.053274] env[61964]: DEBUG nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 743.070668] env[61964]: DEBUG nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 743.070952] env[61964]: DEBUG nova.compute.provider_tree [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 743.090444] env[61964]: DEBUG nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 743.124235] env[61964]: DEBUG nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 743.168624] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280f416-6d7c-9bd1-31b7-e632477cc6bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009733} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.168954] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.169254] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 85cd73df-a2f2-4f54-b984-832f9589a236/85cd73df-a2f2-4f54-b984-832f9589a236.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 743.169545] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d37d602a-228c-411f-8933-0d23b593c16f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.180970] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 743.180970] env[61964]: value = "task-1040796" [ 743.180970] env[61964]: _type = "Task" [ 743.180970] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.193270] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.205866] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.358957} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.207305] env[61964]: DEBUG nova.network.neutron [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updating instance_info_cache with network_info: [{"id": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "address": "fa:16:3e:0a:bb:a5", "network": {"id": "04bdb045-c5b7-45f4-86aa-66cf4b1ab27a", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-487235274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab93cadadb54acbb22a742f37613af3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee1797a0-61", "ovs_interfaceid": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.208538] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 743.209572] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30bf04fb-f5c8-4452-87b4-714934f3e8b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.242073] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 46435c68-f85a-4360-b2b5-6296afc33c3e/46435c68-f85a-4360-b2b5-6296afc33c3e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.246191] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1208b23e-9bf9-47dd-aa8c-a6cb134fb803 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.271106] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 743.271106] env[61964]: value = "task-1040797" [ 743.271106] env[61964]: _type = "Task" [ 743.271106] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.283968] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040797, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.302888] env[61964]: DEBUG nova.compute.manager [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Received event network-vif-plugged-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 743.303741] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Acquiring lock "4104dee9-34db-4396-9189-8dc677d4b677-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.304097] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Lock "4104dee9-34db-4396-9189-8dc677d4b677-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.304344] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Lock "4104dee9-34db-4396-9189-8dc677d4b677-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.304704] env[61964]: DEBUG nova.compute.manager [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] No waiting events found dispatching network-vif-plugged-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 743.304852] env[61964]: WARNING nova.compute.manager [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Received unexpected event network-vif-plugged-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 for instance with vm_state building and task_state spawning. [ 743.305229] env[61964]: DEBUG nova.compute.manager [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Received event network-changed-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 743.305621] env[61964]: DEBUG nova.compute.manager [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Refreshing instance network info cache due to event network-changed-ee1797a0-610e-49f9-9adc-a6ddcb0857a8. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 743.305730] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Acquiring lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.653063] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a908929-21be-4120-a124-28560ca550a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.663909] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb75365f-95cb-4c6d-98a6-3149a9048c6f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.707919] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3ff8ea-3cc1-42bf-872d-c4d2379c8a63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.713264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Releasing lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.713264] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Instance network_info: |[{"id": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "address": "fa:16:3e:0a:bb:a5", "network": {"id": "04bdb045-c5b7-45f4-86aa-66cf4b1ab27a", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-487235274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab93cadadb54acbb22a742f37613af3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee1797a0-61", "ovs_interfaceid": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 743.713836] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Acquired lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.713836] env[61964]: DEBUG nova.network.neutron [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Refreshing network info cache for port ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.714783] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:bb:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee1797a0-610e-49f9-9adc-a6ddcb0857a8', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.723216] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Creating folder: Project (7ab93cadadb54acbb22a742f37613af3). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.728291] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1904880-9ce5-4b4d-b81e-71c11a5c1f4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.731000] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040796, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.735665] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d136b3-50c3-44c4-bf53-622c3cb4197a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.746019] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Created folder: Project (7ab93cadadb54acbb22a742f37613af3) in parent group-v230360. [ 743.746544] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Creating folder: Instances. Parent ref: group-v230442. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.755905] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb26c4b3-66b3-437b-897b-c01765044a1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.758708] env[61964]: DEBUG nova.compute.provider_tree [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 743.771114] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Created folder: Instances in parent group-v230442. [ 743.771315] env[61964]: DEBUG oslo.service.loopingcall [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.771846] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.775809] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b754fba-0811-48c9-9edc-df7c8f4fe02c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.795705] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040797, 'name': ReconfigVM_Task, 'duration_secs': 0.334635} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.797090] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 46435c68-f85a-4360-b2b5-6296afc33c3e/46435c68-f85a-4360-b2b5-6296afc33c3e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.797705] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.797705] env[61964]: value = "task-1040800" [ 743.797705] env[61964]: _type = "Task" [ 743.797705] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.797902] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea380b49-8afd-4eb3-8228-c02428a087cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.808284] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040800, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.809532] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 743.809532] env[61964]: value = "task-1040801" [ 743.809532] env[61964]: _type = "Task" [ 743.809532] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.818237] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040801, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.832875] env[61964]: DEBUG nova.network.neutron [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.213949] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595435} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.214264] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 85cd73df-a2f2-4f54-b984-832f9589a236/85cd73df-a2f2-4f54-b984-832f9589a236.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.214472] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.215400] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf2e4eea-a37a-41c3-a4c5-6c061660a2eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.225712] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 744.225712] env[61964]: value = "task-1040802" [ 744.225712] env[61964]: _type = "Task" [ 744.225712] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.234811] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040802, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.300187] env[61964]: DEBUG nova.scheduler.client.report [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 744.300474] env[61964]: DEBUG nova.compute.provider_tree [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 73 to 74 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 744.300657] env[61964]: DEBUG nova.compute.provider_tree [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 744.316657] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040800, 'name': CreateVM_Task, 'duration_secs': 0.318073} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.317205] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 744.317923] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.318109] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.318448] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.318699] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c2c744f-f8cc-4688-882d-09c239e980f5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.323969] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040801, 'name': Rename_Task, 'duration_secs': 0.148685} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.325739] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 744.326323] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b1a56d8-982d-4d94-81a0-9d80e6561a86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.329749] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 744.329749] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524ef6d9-49d1-dc77-e0c2-5ffa4818e15f" [ 744.329749] env[61964]: _type = "Task" [ 744.329749] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.337860] env[61964]: INFO nova.compute.manager [-] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Took 1.49 seconds to deallocate network for instance. [ 744.338317] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 744.338317] env[61964]: value = "task-1040803" [ 744.338317] env[61964]: _type = "Task" [ 744.338317] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.350620] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524ef6d9-49d1-dc77-e0c2-5ffa4818e15f, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.355207] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.359119] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.359119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.359119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.359119] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.359119] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0751fa13-ab43-435d-bca0-6de24d730cec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.363548] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040803, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.375364] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.376766] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.376766] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ec15e32-b14a-4d4b-858c-c0dd76d6bb1e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.383631] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 744.383631] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dec1b9-396d-6166-5d79-68db2add45d7" [ 744.383631] env[61964]: _type = "Task" [ 744.383631] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.395131] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dec1b9-396d-6166-5d79-68db2add45d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.517417] env[61964]: DEBUG nova.network.neutron [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updated VIF entry in instance network info cache for port ee1797a0-610e-49f9-9adc-a6ddcb0857a8. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.517832] env[61964]: DEBUG nova.network.neutron [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updating instance_info_cache with network_info: [{"id": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "address": "fa:16:3e:0a:bb:a5", "network": {"id": "04bdb045-c5b7-45f4-86aa-66cf4b1ab27a", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-487235274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab93cadadb54acbb22a742f37613af3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee1797a0-61", "ovs_interfaceid": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.736458] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040802, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110864} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.736802] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.737668] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd0badb-fe8a-4f8c-ba69-2f2f9724542a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.763039] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 85cd73df-a2f2-4f54-b984-832f9589a236/85cd73df-a2f2-4f54-b984-832f9589a236.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.763284] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71470628-7ffb-427c-bfc4-4666af3e76a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.785165] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 744.785165] env[61964]: value = "task-1040804" [ 744.785165] env[61964]: _type = "Task" [ 744.785165] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.797430] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040804, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.808497] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.893s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.809061] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.811975] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.218s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.812722] env[61964]: DEBUG nova.objects.instance [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lazy-loading 'resources' on Instance uuid 5c340420-969b-4ccb-9f9a-7a833d8f0c43 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 744.855366] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040803, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.858331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.894701] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dec1b9-396d-6166-5d79-68db2add45d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010583} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.895654] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e99910e-901a-4b77-a5da-0402029ab8e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.901815] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 744.901815] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52460d94-127a-4f85-7876-9ed7fb9e7c4b" [ 744.901815] env[61964]: _type = "Task" [ 744.901815] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.910157] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52460d94-127a-4f85-7876-9ed7fb9e7c4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.020296] env[61964]: DEBUG oslo_concurrency.lockutils [req-f1772d90-505d-429c-a9ca-d97b2b9f8208 req-852f952e-e718-4e55-9313-5cdf033abe70 service nova] Releasing lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.295102] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040804, 'name': ReconfigVM_Task, 'duration_secs': 0.304577} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.295418] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 85cd73df-a2f2-4f54-b984-832f9589a236/85cd73df-a2f2-4f54-b984-832f9589a236.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.295985] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c782a674-0ab6-4c76-acdc-0a14b3be9569 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.302595] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 745.302595] env[61964]: value = "task-1040805" [ 745.302595] env[61964]: _type = "Task" [ 745.302595] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.310898] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040805, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.315261] env[61964]: DEBUG nova.compute.utils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.319626] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.319731] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 745.329217] env[61964]: DEBUG nova.compute.manager [req-35109dae-9678-4906-b2a9-b3af811a41c3 req-261f7107-0aad-452f-b164-bf69dea1e3d9 service nova] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Received event network-vif-deleted-b5773bca-cae3-4d80-9784-11117858828f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 745.352009] env[61964]: DEBUG oslo_vmware.api [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040803, 'name': PowerOnVM_Task, 'duration_secs': 0.749517} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.352343] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 745.352608] env[61964]: INFO nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Took 9.64 seconds to spawn the instance on the hypervisor. [ 745.353159] env[61964]: DEBUG nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 745.355796] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6120d3bf-7c7a-41ce-881d-98b22bfcd908 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.368260] env[61964]: DEBUG nova.policy [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '074d07c11c994a1e9183f520daf35654', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c44820ae8513473cbbdfddb36877a9b3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.415527] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52460d94-127a-4f85-7876-9ed7fb9e7c4b, 'name': SearchDatastore_Task, 'duration_secs': 0.010386} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.418651] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.419799] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4104dee9-34db-4396-9189-8dc677d4b677/4104dee9-34db-4396-9189-8dc677d4b677.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 745.419799] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3111315d-e94b-4d90-beee-c79d3b9c41dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.441446] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 745.441446] env[61964]: value = "task-1040806" [ 745.441446] env[61964]: _type = "Task" [ 745.441446] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.453311] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.774935] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfc0faf-f29d-41cc-8cf8-bb3c8e632405 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.785103] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c8f97c-28aa-4d3f-8b79-a0929dfa5725 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.824337] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.832471] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c3e1aa-0505-4894-9e92-bf778247ba69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.842996] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040805, 'name': Rename_Task, 'duration_secs': 0.155494} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.847132] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 745.847132] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d266f06-0dbf-4374-94fe-d6b85aefd287 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.848589] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aedbc69-2b51-479c-8395-a3c3bd207304 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.872168] env[61964]: DEBUG nova.compute.provider_tree [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.872736] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Successfully created port: 3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.879189] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 745.879189] env[61964]: value = "task-1040807" [ 745.879189] env[61964]: _type = "Task" [ 745.879189] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.879699] env[61964]: INFO nova.compute.manager [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Took 38.72 seconds to build instance. [ 745.895177] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040807, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.953964] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040806, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.381865] env[61964]: DEBUG nova.scheduler.client.report [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 746.385365] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48ab47a7-bdfe-48a7-aefb-a68a4495595e tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.711s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.398545] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040807, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.454464] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545742} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.454758] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4104dee9-34db-4396-9189-8dc677d4b677/4104dee9-34db-4396-9189-8dc677d4b677.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 746.454974] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.455281] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c04b33c3-0270-404e-89bf-132dbdadeb17 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.463722] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 746.463722] env[61964]: value = "task-1040808" [ 746.463722] env[61964]: _type = "Task" [ 746.463722] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.472887] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.843182] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.869595] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 746.869737] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.869888] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 746.870087] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.870244] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 746.870400] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 746.870617] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 746.870783] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 746.870954] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 746.871187] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 746.871316] env[61964]: DEBUG nova.virt.hardware [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 746.872189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9585ff-0cc0-4ff2-8083-3d029c838ba8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.882428] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c4751b-e5fe-473e-ad39-c3e256de997f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.886669] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.892813] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.677s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.894435] env[61964]: INFO nova.compute.claims [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.900017] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 746.913740] env[61964]: INFO nova.scheduler.client.report [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Deleted allocations for instance 5c340420-969b-4ccb-9f9a-7a833d8f0c43 [ 746.917917] env[61964]: DEBUG oslo_vmware.api [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040807, 'name': PowerOnVM_Task, 'duration_secs': 0.615595} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.917917] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 746.917917] env[61964]: INFO nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Took 8.63 seconds to spawn the instance on the hypervisor. [ 746.917917] env[61964]: DEBUG nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 746.921543] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece33b6f-a66d-4220-bef6-de9bf112cbc2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.975280] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068481} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.975649] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.976810] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b289c576-e250-4490-b2c1-62492380299f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.001855] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 4104dee9-34db-4396-9189-8dc677d4b677/4104dee9-34db-4396-9189-8dc677d4b677.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.002418] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0631d13-bdca-46a2-b133-1a6d19dbb446 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.023913] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 747.023913] env[61964]: value = "task-1040809" [ 747.023913] env[61964]: _type = "Task" [ 747.023913] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.033163] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.428623] env[61964]: DEBUG oslo_concurrency.lockutils [None req-417675d9-40be-429e-852e-3d56302f9983 tempest-ServerAddressesTestJSON-1886251295 tempest-ServerAddressesTestJSON-1886251295-project-member] Lock "5c340420-969b-4ccb-9f9a-7a833d8f0c43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.342s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.438245] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.440749] env[61964]: INFO nova.compute.manager [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Took 36.78 seconds to build instance. [ 747.499095] env[61964]: DEBUG nova.compute.manager [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Received event network-vif-plugged-3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 747.499331] env[61964]: DEBUG oslo_concurrency.lockutils [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] Acquiring lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.499901] env[61964]: DEBUG oslo_concurrency.lockutils [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.499901] env[61964]: DEBUG oslo_concurrency.lockutils [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.499901] env[61964]: DEBUG nova.compute.manager [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] No waiting events found dispatching network-vif-plugged-3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 747.500129] env[61964]: WARNING nova.compute.manager [req-f00bd0be-c858-4f64-bd33-220308c84f81 req-e86fb461-1285-4b4f-9f87-aa1e21c09e49 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Received unexpected event network-vif-plugged-3038d60e-1ce8-479d-ac96-8b597d6d8d00 for instance with vm_state building and task_state spawning. [ 747.534929] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040809, 'name': ReconfigVM_Task, 'duration_secs': 0.332016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.535197] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 4104dee9-34db-4396-9189-8dc677d4b677/4104dee9-34db-4396-9189-8dc677d4b677.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.535845] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7735496e-813e-4acd-9dfb-a99e24e9aec1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.543774] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 747.543774] env[61964]: value = "task-1040810" [ 747.543774] env[61964]: _type = "Task" [ 747.543774] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.552461] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040810, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.553611] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Successfully updated port: 3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 747.945868] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4ff7711d-f036-4a91-a76c-8247bccd2352 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.091s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.058382] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.058514] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquired lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.058735] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.060458] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040810, 'name': Rename_Task, 'duration_secs': 0.158394} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.060888] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 748.061217] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-308aec27-ca39-4e8c-a025-e40415cfc9b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.070534] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 748.070534] env[61964]: value = "task-1040811" [ 748.070534] env[61964]: _type = "Task" [ 748.070534] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.084438] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.385604] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.385732] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.405595] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58587752-0a03-4be8-ad7a-23642c9ade3a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.414533] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70b7664-2730-4964-9736-342a370157cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.451463] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 748.456504] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a993924-79bd-4be3-8d0b-89884490a3c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.465430] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce9a4be-2d48-4bd1-9e98-ddbede67c65c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.486972] env[61964]: DEBUG nova.compute.provider_tree [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.583161] env[61964]: DEBUG oslo_vmware.api [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040811, 'name': PowerOnVM_Task, 'duration_secs': 0.487648} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.583161] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.583161] env[61964]: INFO nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Took 7.70 seconds to spawn the instance on the hypervisor. [ 748.583161] env[61964]: DEBUG nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 748.583740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3825759-3a14-466b-a325-028771764a46 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.807320] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.983397] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.992250] env[61964]: DEBUG nova.scheduler.client.report [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.014046] env[61964]: DEBUG nova.network.neutron [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updating instance_info_cache with network_info: [{"id": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "address": "fa:16:3e:60:92:36", "network": {"id": "cf9d27f7-1266-4cd4-8407-05696aa41712", "bridge": "br-int", "label": "tempest-ServersTestJSON-981760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c44820ae8513473cbbdfddb36877a9b3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3038d60e-1c", "ovs_interfaceid": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.041386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "85cd73df-a2f2-4f54-b984-832f9589a236" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.041724] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.042016] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.042268] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.042466] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.045471] env[61964]: INFO nova.compute.manager [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Terminating instance [ 749.104956] env[61964]: INFO nova.compute.manager [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Took 35.96 seconds to build instance. [ 749.497335] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.498014] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 749.500557] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.334s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.501993] env[61964]: INFO nova.compute.claims [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.517616] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Releasing lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.517980] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Instance network_info: |[{"id": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "address": "fa:16:3e:60:92:36", "network": {"id": "cf9d27f7-1266-4cd4-8407-05696aa41712", "bridge": "br-int", "label": "tempest-ServersTestJSON-981760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c44820ae8513473cbbdfddb36877a9b3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3038d60e-1c", "ovs_interfaceid": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 749.518453] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:92:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '41f66e20-fd86-4158-bbdc-7a150e85e844', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3038d60e-1ce8-479d-ac96-8b597d6d8d00', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.531146] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Creating folder: Project (c44820ae8513473cbbdfddb36877a9b3). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.531623] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9901809c-20f1-457a-b6a1-b41100c0f4e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.548021] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Created folder: Project (c44820ae8513473cbbdfddb36877a9b3) in parent group-v230360. [ 749.548021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Creating folder: Instances. Parent ref: group-v230445. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.548021] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fd13ca4-a0ac-4b7c-990d-97505393f7bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.553580] env[61964]: DEBUG nova.compute.manager [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 749.553968] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 749.556039] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d321e54-7c54-46c0-aba3-5bbaa617622b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.567787] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 749.568253] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d587f40-df3d-4bef-9122-eb972a7d493d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.573719] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Created folder: Instances in parent group-v230445. [ 749.574127] env[61964]: DEBUG oslo.service.loopingcall [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.574345] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 749.574776] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce3d72d8-f56b-49ae-8e1f-96855d4dc677 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.595037] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 749.595037] env[61964]: value = "task-1040814" [ 749.595037] env[61964]: _type = "Task" [ 749.595037] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.603482] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.603482] env[61964]: value = "task-1040815" [ 749.603482] env[61964]: _type = "Task" [ 749.603482] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.607159] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.611045] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bf598652-322e-4706-86a8-d6e10b747196 tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.737s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.621109] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040815, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.642568] env[61964]: DEBUG nova.compute.manager [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Received event network-changed-3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 749.642568] env[61964]: DEBUG nova.compute.manager [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Refreshing instance network info cache due to event network-changed-3038d60e-1ce8-479d-ac96-8b597d6d8d00. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 749.642568] env[61964]: DEBUG oslo_concurrency.lockutils [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] Acquiring lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.642568] env[61964]: DEBUG oslo_concurrency.lockutils [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] Acquired lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.642568] env[61964]: DEBUG nova.network.neutron [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Refreshing network info cache for port 3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 750.006525] env[61964]: DEBUG nova.compute.utils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.008535] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 750.012258] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 750.088196] env[61964]: DEBUG nova.policy [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '32a91d6374524d48bade7765b5691865', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1899a050952647d390b369ca6aeb2156', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.106501] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040814, 'name': PowerOffVM_Task, 'duration_secs': 0.224145} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.107335] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 750.107815] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 750.113382] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd89aeb1-cea3-459d-b3f9-e5324a4a0195 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.117029] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 750.125500] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040815, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.187029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 750.187029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 750.187029] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Deleting the datastore file [datastore2] 85cd73df-a2f2-4f54-b984-832f9589a236 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.187307] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da97a475-817f-413f-b434-a65876d4e1b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.198510] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for the task: (returnval){ [ 750.198510] env[61964]: value = "task-1040817" [ 750.198510] env[61964]: _type = "Task" [ 750.198510] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.209935] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.518601] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 750.568087] env[61964]: DEBUG nova.network.neutron [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updated VIF entry in instance network info cache for port 3038d60e-1ce8-479d-ac96-8b597d6d8d00. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 750.568638] env[61964]: DEBUG nova.network.neutron [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updating instance_info_cache with network_info: [{"id": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "address": "fa:16:3e:60:92:36", "network": {"id": "cf9d27f7-1266-4cd4-8407-05696aa41712", "bridge": "br-int", "label": "tempest-ServersTestJSON-981760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c44820ae8513473cbbdfddb36877a9b3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3038d60e-1c", "ovs_interfaceid": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.619024] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Successfully created port: 307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.622874] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040815, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.644495] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.717309] env[61964]: DEBUG oslo_vmware.api [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Task: {'id': task-1040817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143769} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.717900] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 750.717900] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 750.718038] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 750.718135] env[61964]: INFO nova.compute.manager [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Took 1.16 seconds to destroy the instance on the hypervisor. [ 750.718386] env[61964]: DEBUG oslo.service.loopingcall [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.718754] env[61964]: DEBUG nova.compute.manager [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 750.718754] env[61964]: DEBUG nova.network.neutron [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 750.957771] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7d5a55-4ff4-4a9e-b67f-9341b5643f15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.966387] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c88a5d-9e8f-4ece-8768-c07b7d681418 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.000801] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ce42d2-72e4-4523-9b58-2cced02a292f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.009791] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8dc889-55db-433e-b8a7-dc852fe6b239 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.034989] env[61964]: DEBUG nova.compute.provider_tree [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.082132] env[61964]: DEBUG oslo_concurrency.lockutils [req-0068f7ae-a569-4203-a34a-5ad017211dc5 req-21e40175-a64b-486c-a2f6-ca9c55dc0bd6 service nova] Releasing lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.120307] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040815, 'name': CreateVM_Task, 'duration_secs': 1.466408} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.120676] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 751.121456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.121749] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.122068] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.122517] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90179333-38b8-4e35-9f7c-a3c4561be076 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.128324] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 751.128324] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ec35d0-e2c5-1732-8df0-dd9d88cab797" [ 751.128324] env[61964]: _type = "Task" [ 751.128324] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.137874] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ec35d0-e2c5-1732-8df0-dd9d88cab797, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.538612] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 751.541601] env[61964]: DEBUG nova.scheduler.client.report [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 751.567644] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 751.567939] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.568148] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 751.568373] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.568544] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 751.568700] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 751.568912] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 751.569124] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 751.569274] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 751.569484] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 751.569677] env[61964]: DEBUG nova.virt.hardware [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 751.570534] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713107ea-e2f3-43d0-9761-89f5d02fe292 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.579888] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2537ad9d-aeb0-4042-ab9c-c9d5428be86e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.639055] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ec35d0-e2c5-1732-8df0-dd9d88cab797, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.639369] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.639615] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 751.639852] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.640013] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.640203] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 751.640465] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e9236d0-0797-4cb0-afa4-0c8f95764fd6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.648729] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.648912] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 751.649652] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8c06aec-a1c4-4b0e-9d05-d570df654086 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.655016] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 751.655016] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b3def-0f4a-ab65-0f2f-c1170ed9aa0a" [ 751.655016] env[61964]: _type = "Task" [ 751.655016] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.663014] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b3def-0f4a-ab65-0f2f-c1170ed9aa0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.671645] env[61964]: DEBUG nova.compute.manager [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Received event network-changed-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 751.671840] env[61964]: DEBUG nova.compute.manager [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Refreshing instance network info cache due to event network-changed-ee1797a0-610e-49f9-9adc-a6ddcb0857a8. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 751.672079] env[61964]: DEBUG oslo_concurrency.lockutils [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] Acquiring lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.672234] env[61964]: DEBUG oslo_concurrency.lockutils [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] Acquired lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.672396] env[61964]: DEBUG nova.network.neutron [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Refreshing network info cache for port ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.673832] env[61964]: DEBUG nova.network.neutron [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.047362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.048112] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 752.052134] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.114s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.054101] env[61964]: INFO nova.compute.claims [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.170964] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526b3def-0f4a-ab65-0f2f-c1170ed9aa0a, 'name': SearchDatastore_Task, 'duration_secs': 0.010295} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.171823] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84fbbdea-1f9b-4863-9911-ad0dc5733a91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.178057] env[61964]: INFO nova.compute.manager [-] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Took 1.46 seconds to deallocate network for instance. [ 752.181163] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 752.181163] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52489c05-1aee-930d-e7d8-1789ca2d6bb0" [ 752.181163] env[61964]: _type = "Task" [ 752.181163] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.194440] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52489c05-1aee-930d-e7d8-1789ca2d6bb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.547805] env[61964]: DEBUG nova.network.neutron [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updated VIF entry in instance network info cache for port ee1797a0-610e-49f9-9adc-a6ddcb0857a8. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.548345] env[61964]: DEBUG nova.network.neutron [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updating instance_info_cache with network_info: [{"id": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "address": "fa:16:3e:0a:bb:a5", "network": {"id": "04bdb045-c5b7-45f4-86aa-66cf4b1ab27a", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-487235274-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab93cadadb54acbb22a742f37613af3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee1797a0-61", "ovs_interfaceid": "ee1797a0-610e-49f9-9adc-a6ddcb0857a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.559242] env[61964]: DEBUG nova.compute.utils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.565996] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 752.565996] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 752.632177] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Successfully updated port: 307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.651900] env[61964]: DEBUG nova.policy [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '731ab2c060944f41b8bd293b3282a703', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912fe686053f4633b1099dce01d72365', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.690572] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.694110] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52489c05-1aee-930d-e7d8-1789ca2d6bb0, 'name': SearchDatastore_Task, 'duration_secs': 0.034201} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.694110] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.694110] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 58f8d368-fa0d-4c8d-9009-e6e1c5940032/58f8d368-fa0d-4c8d-9009-e6e1c5940032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 752.694110] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09f3138b-65a9-41f1-a4a3-ea42db1fb6e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.701740] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 752.701740] env[61964]: value = "task-1040818" [ 752.701740] env[61964]: _type = "Task" [ 752.701740] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.710629] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040818, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.964566] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Successfully created port: df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 753.052271] env[61964]: DEBUG oslo_concurrency.lockutils [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] Releasing lock "refresh_cache-4104dee9-34db-4396-9189-8dc677d4b677" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.052584] env[61964]: DEBUG nova.compute.manager [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Received event network-vif-deleted-b94b591f-9808-4308-b7b6-739f0b48f4fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 753.053122] env[61964]: INFO nova.compute.manager [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Neutron deleted interface b94b591f-9808-4308-b7b6-739f0b48f4fa; detaching it from the instance and deleting it from the info cache [ 753.053462] env[61964]: DEBUG nova.network.neutron [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.064528] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 753.134481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.134659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquired lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.134784] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.216809] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040818, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.334551] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Successfully created port: 3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 753.540551] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b7d0f7-2fcd-42e2-b861-0516eb599884 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.551111] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a1bae3-dcef-4179-be8b-e96e84af80ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.556300] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-227ddf1a-2723-4642-888f-7269a1e72026 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.586777] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b11f97-df7f-40de-bf57-d7f6dc6b5b1f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.593538] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3966d18e-0640-461f-83d1-8e6ead80910a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.607238] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8599cd-7b3b-4189-8474-c6c9f1240270 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.622452] env[61964]: DEBUG nova.compute.provider_tree [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 753.635658] env[61964]: DEBUG nova.compute.manager [req-d06b3f92-a81e-49a1-8252-87e2fc5eda8d req-3f8e332a-5d30-4ee9-bd55-1d00d46fcea5 service nova] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Detach interface failed, port_id=b94b591f-9808-4308-b7b6-739f0b48f4fa, reason: Instance 85cd73df-a2f2-4f54-b984-832f9589a236 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 753.665464] env[61964]: ERROR nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [req-dbfc2784-b01a-4ace-a850-18ed4d6c4ba2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-dbfc2784-b01a-4ace-a850-18ed4d6c4ba2"}]} [ 753.688918] env[61964]: DEBUG nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 753.700374] env[61964]: DEBUG nova.compute.manager [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Received event network-vif-plugged-307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 753.700601] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Acquiring lock "fb538d67-47c3-4639-a642-b3358f97df31-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.700811] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Lock "fb538d67-47c3-4639-a642-b3358f97df31-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.700983] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Lock "fb538d67-47c3-4639-a642-b3358f97df31-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.701168] env[61964]: DEBUG nova.compute.manager [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] No waiting events found dispatching network-vif-plugged-307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 753.701338] env[61964]: WARNING nova.compute.manager [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Received unexpected event network-vif-plugged-307bc2ba-d780-4df4-a420-4aa33275a001 for instance with vm_state building and task_state spawning. [ 753.701500] env[61964]: DEBUG nova.compute.manager [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Received event network-changed-307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 753.701656] env[61964]: DEBUG nova.compute.manager [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Refreshing instance network info cache due to event network-changed-307bc2ba-d780-4df4-a420-4aa33275a001. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 753.701824] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Acquiring lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.702593] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.705671] env[61964]: DEBUG nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 753.705871] env[61964]: DEBUG nova.compute.provider_tree [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 753.717161] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040818, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597668} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.717413] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 58f8d368-fa0d-4c8d-9009-e6e1c5940032/58f8d368-fa0d-4c8d-9009-e6e1c5940032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 753.717627] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.717870] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b0d16b7-54a0-454d-b8c2-05cb12502eac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.724220] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 753.724220] env[61964]: value = "task-1040819" [ 753.724220] env[61964]: _type = "Task" [ 753.724220] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.725021] env[61964]: DEBUG nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 753.737263] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.744338] env[61964]: DEBUG nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 753.854857] env[61964]: DEBUG nova.network.neutron [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Updating instance_info_cache with network_info: [{"id": "307bc2ba-d780-4df4-a420-4aa33275a001", "address": "fa:16:3e:7b:10:eb", "network": {"id": "f1b7f518-bf79-46c5-95e0-1d354a583c2e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1168020866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1899a050952647d390b369ca6aeb2156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f1e0e39-0c84-4fcd-9113-cc528c3eb185", "external-id": "nsx-vlan-transportzone-907", "segmentation_id": 907, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307bc2ba-d7", "ovs_interfaceid": "307bc2ba-d780-4df4-a420-4aa33275a001", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.090832] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 754.103500] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121b18c0-83fb-4eb7-a8b1-fcf3e826c5c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.116183] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983ee781-3fdf-452b-a3de-d2c8bbda67cc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.121100] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 754.121386] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.121687] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 754.121887] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.122053] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 754.122213] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 754.122441] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 754.122612] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 754.122783] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 754.122949] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 754.123135] env[61964]: DEBUG nova.virt.hardware [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 754.123978] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dc31aa-7a25-4f7e-813d-c12653dd2d42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.152854] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c8480a-58f1-423c-9f29-77ffa69c5a43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.159682] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd211ca8-7fc6-4d1e-b325-d57b8408e943 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.167185] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7a88d0-1441-4f31-8179-eac326682a83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.190660] env[61964]: DEBUG nova.compute.provider_tree [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 754.237887] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091105} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.237887] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 754.238191] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc947faf-4bac-4920-a777-6c4e58934521 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.260715] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 58f8d368-fa0d-4c8d-9009-e6e1c5940032/58f8d368-fa0d-4c8d-9009-e6e1c5940032.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.260835] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b5409dd-42a7-4487-8490-7bb5bfc7a4f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.282461] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 754.282461] env[61964]: value = "task-1040820" [ 754.282461] env[61964]: _type = "Task" [ 754.282461] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.291237] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.358468] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Releasing lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.358921] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Instance network_info: |[{"id": "307bc2ba-d780-4df4-a420-4aa33275a001", "address": "fa:16:3e:7b:10:eb", "network": {"id": "f1b7f518-bf79-46c5-95e0-1d354a583c2e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1168020866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1899a050952647d390b369ca6aeb2156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f1e0e39-0c84-4fcd-9113-cc528c3eb185", "external-id": "nsx-vlan-transportzone-907", "segmentation_id": 907, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307bc2ba-d7", "ovs_interfaceid": "307bc2ba-d780-4df4-a420-4aa33275a001", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 754.359160] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Acquired lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.359362] env[61964]: DEBUG nova.network.neutron [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Refreshing network info cache for port 307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.360606] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:10:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f1e0e39-0c84-4fcd-9113-cc528c3eb185', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '307bc2ba-d780-4df4-a420-4aa33275a001', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.368268] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Creating folder: Project (1899a050952647d390b369ca6aeb2156). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.368820] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-524d5ba2-7cb6-494f-aeae-60c6742b4786 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.382389] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Created folder: Project (1899a050952647d390b369ca6aeb2156) in parent group-v230360. [ 754.382620] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Creating folder: Instances. Parent ref: group-v230448. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 754.382839] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cdc4073-d94d-49ef-b370-cb75437f0127 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.393249] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Created folder: Instances in parent group-v230448. [ 754.393592] env[61964]: DEBUG oslo.service.loopingcall [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.393691] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 754.393906] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eedc6aac-df75-40fc-8d18-93dcfe8c67f5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.414250] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.414250] env[61964]: value = "task-1040823" [ 754.414250] env[61964]: _type = "Task" [ 754.414250] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.422722] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040823, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.733834] env[61964]: DEBUG nova.scheduler.client.report [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 75 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 754.734142] env[61964]: DEBUG nova.compute.provider_tree [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 75 to 76 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 754.734338] env[61964]: DEBUG nova.compute.provider_tree [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 754.793273] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040820, 'name': ReconfigVM_Task, 'duration_secs': 0.438103} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.793578] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 58f8d368-fa0d-4c8d-9009-e6e1c5940032/58f8d368-fa0d-4c8d-9009-e6e1c5940032.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.794305] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-299d784a-1412-42b0-b5c9-6d15bd739844 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.801380] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 754.801380] env[61964]: value = "task-1040824" [ 754.801380] env[61964]: _type = "Task" [ 754.801380] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.811304] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040824, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.927473] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040823, 'name': CreateVM_Task, 'duration_secs': 0.364527} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.927934] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 754.932034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.932034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.932034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 754.932034] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceb2ce25-9936-48ae-abfa-8713c291d67a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.936807] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 754.936807] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522ccec4-3060-dd09-aeac-7d8e3779fa11" [ 754.936807] env[61964]: _type = "Task" [ 754.936807] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.947719] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522ccec4-3060-dd09-aeac-7d8e3779fa11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.173942] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Successfully updated port: df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 755.243093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.191s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.243877] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.246559] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 25.526s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.246761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.246917] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 755.247263] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.083s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.249126] env[61964]: INFO nova.compute.claims [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.256025] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6e23e4-9669-4717-9635-63a37195b620 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.268119] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ea7f69-1339-4022-9ad7-155fa12b6f1f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.285887] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c77ade3-581b-4018-b013-a4bee7071e0f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.295329] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e08ad16-f9be-4fc9-9218-3835890c24cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.334758] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180429MB free_disk=182GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 755.335772] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.341458] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040824, 'name': Rename_Task, 'duration_secs': 0.15251} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.341743] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 755.341995] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebd28022-b230-454d-bca9-cd249484ea58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.350084] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 755.350084] env[61964]: value = "task-1040825" [ 755.350084] env[61964]: _type = "Task" [ 755.350084] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.358510] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.441263] env[61964]: DEBUG nova.network.neutron [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Updated VIF entry in instance network info cache for port 307bc2ba-d780-4df4-a420-4aa33275a001. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 755.441656] env[61964]: DEBUG nova.network.neutron [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Updating instance_info_cache with network_info: [{"id": "307bc2ba-d780-4df4-a420-4aa33275a001", "address": "fa:16:3e:7b:10:eb", "network": {"id": "f1b7f518-bf79-46c5-95e0-1d354a583c2e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1168020866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1899a050952647d390b369ca6aeb2156", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f1e0e39-0c84-4fcd-9113-cc528c3eb185", "external-id": "nsx-vlan-transportzone-907", "segmentation_id": 907, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap307bc2ba-d7", "ovs_interfaceid": "307bc2ba-d780-4df4-a420-4aa33275a001", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.453337] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522ccec4-3060-dd09-aeac-7d8e3779fa11, 'name': SearchDatastore_Task, 'duration_secs': 0.012961} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.454237] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.454565] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.454708] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.454861] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.455056] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.455948] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-89bc8088-dbc7-4e76-b808-9d5df69a96e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.471020] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.471020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 755.471020] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af54e272-c943-4b0d-a2c6-91c0510f21d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.479836] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 755.479836] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dc2d87-f702-4232-ffbc-8b1e72aaba74" [ 755.479836] env[61964]: _type = "Task" [ 755.479836] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.490844] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dc2d87-f702-4232-ffbc-8b1e72aaba74, 'name': SearchDatastore_Task, 'duration_secs': 0.009292} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.494201] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62481de5-bdbd-493c-bb9c-4230c02406fb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.500086] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 755.500086] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526e84ad-d334-b798-2162-9a0965efb954" [ 755.500086] env[61964]: _type = "Task" [ 755.500086] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.511314] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526e84ad-d334-b798-2162-9a0965efb954, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.723081] env[61964]: DEBUG nova.compute.manager [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-vif-plugged-df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 755.723335] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Acquiring lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.723635] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.723787] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.723986] env[61964]: DEBUG nova.compute.manager [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] No waiting events found dispatching network-vif-plugged-df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 755.724199] env[61964]: WARNING nova.compute.manager [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received unexpected event network-vif-plugged-df5c5bf6-0c91-49b4-b091-0fa57f37ec46 for instance with vm_state building and task_state spawning. [ 755.724449] env[61964]: DEBUG nova.compute.manager [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-changed-df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 755.724764] env[61964]: DEBUG nova.compute.manager [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Refreshing instance network info cache due to event network-changed-df5c5bf6-0c91-49b4-b091-0fa57f37ec46. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 755.724830] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Acquiring lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.724982] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Acquired lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.725160] env[61964]: DEBUG nova.network.neutron [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Refreshing network info cache for port df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 755.756222] env[61964]: DEBUG nova.compute.utils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.758418] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 755.758803] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.844532] env[61964]: DEBUG nova.policy [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '771d16e2520d4c2e8161ba3da91a03ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19bac13e505143b4ae5884c24c748b1a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.863036] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040825, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.947243] env[61964]: DEBUG oslo_concurrency.lockutils [req-e105e195-1a8f-45fc-acd0-e3ac4460096f req-f9defcfb-76df-42e3-9989-11e5cea8ebcd service nova] Releasing lock "refresh_cache-fb538d67-47c3-4639-a642-b3358f97df31" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.013597] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526e84ad-d334-b798-2162-9a0965efb954, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.013892] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.014175] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] fb538d67-47c3-4639-a642-b3358f97df31/fb538d67-47c3-4639-a642-b3358f97df31.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 756.014564] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30fbfccf-0b45-4e18-bc14-bef63c3758bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.022485] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 756.022485] env[61964]: value = "task-1040826" [ 756.022485] env[61964]: _type = "Task" [ 756.022485] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.034203] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040826, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.268421] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.276246] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Successfully created port: 1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.278432] env[61964]: DEBUG nova.network.neutron [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.367651] env[61964]: DEBUG oslo_vmware.api [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040825, 'name': PowerOnVM_Task, 'duration_secs': 0.611572} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.371302] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 756.371596] env[61964]: INFO nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Took 9.53 seconds to spawn the instance on the hypervisor. [ 756.371801] env[61964]: DEBUG nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 756.373231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6dd5a4-2dda-43a5-806c-f90bfc12ca97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.486539] env[61964]: DEBUG nova.network.neutron [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.534314] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040826, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.758946] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3d168b-46aa-4acf-abd6-3e5d72e1bfd2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.769419] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f3ed9c-4b5f-4766-9dc6-7d655e2c2583 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.802444] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e404be1-3ae5-4da8-aa56-346cda330b14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.811806] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556c61ad-421f-4a0a-9164-ec2b0a187add {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.826560] env[61964]: DEBUG nova.compute.provider_tree [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.905603] env[61964]: INFO nova.compute.manager [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Took 36.78 seconds to build instance. [ 756.992527] env[61964]: DEBUG oslo_concurrency.lockutils [req-7f41facd-1fb2-4717-9dd6-a1b9b0a426e1 req-0e23c61d-793d-41cb-a151-f0fab2cdeb3b service nova] Releasing lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.034575] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040826, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528219} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.034777] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] fb538d67-47c3-4639-a642-b3358f97df31/fb538d67-47c3-4639-a642-b3358f97df31.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 757.034996] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.035275] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcae0548-147f-47d7-987f-33cfec632dba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.045195] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 757.045195] env[61964]: value = "task-1040827" [ 757.045195] env[61964]: _type = "Task" [ 757.045195] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.054718] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040827, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.304019] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.330353] env[61964]: DEBUG nova.scheduler.client.report [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 757.338233] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 757.338463] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.338652] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.338864] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.339054] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.339216] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 757.339429] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 757.339599] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 757.339946] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 757.339946] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 757.340119] env[61964]: DEBUG nova.virt.hardware [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 757.340998] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e16ad7-2c24-4959-9f51-bb02e1aa44a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.350816] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dd3706-9f0f-43bd-90bd-940409e92c8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.408536] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Successfully updated port: 3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 757.410157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0383219-23d1-4c91-a73d-fc93dd3f3cfc tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.664s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.555730] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040827, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.208432} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.556538] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.556892] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7a321e-9753-4d98-982d-b165ed0dbc44 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.580078] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] fb538d67-47c3-4639-a642-b3358f97df31/fb538d67-47c3-4639-a642-b3358f97df31.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.580388] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6febdbc6-00fd-488b-bf68-611251ef3d28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.601033] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 757.601033] env[61964]: value = "task-1040828" [ 757.601033] env[61964]: _type = "Task" [ 757.601033] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.610010] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040828, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.788198] env[61964]: DEBUG nova.compute.manager [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-vif-plugged-3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 757.788198] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Acquiring lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.788198] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.788198] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.788198] env[61964]: DEBUG nova.compute.manager [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] No waiting events found dispatching network-vif-plugged-3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 757.788572] env[61964]: WARNING nova.compute.manager [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received unexpected event network-vif-plugged-3e1917c9-a522-4ee0-9a2c-7808a7918985 for instance with vm_state building and task_state spawning. [ 757.788572] env[61964]: DEBUG nova.compute.manager [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-changed-3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 757.789138] env[61964]: DEBUG nova.compute.manager [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Refreshing instance network info cache due to event network-changed-3e1917c9-a522-4ee0-9a2c-7808a7918985. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 757.789628] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Acquiring lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.789906] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Acquired lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.791540] env[61964]: DEBUG nova.network.neutron [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Refreshing network info cache for port 3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.845659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.846282] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 757.849517] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.640s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.850013] env[61964]: DEBUG nova.objects.instance [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 757.912053] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.915671] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 758.048321] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Successfully updated port: 1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 758.118439] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040828, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.324958] env[61964]: DEBUG nova.network.neutron [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.357963] env[61964]: DEBUG nova.compute.utils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.359357] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 758.359531] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 758.425048] env[61964]: DEBUG nova.policy [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05bfff4e5dae488593812635d1c216b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f1804a0ded74317a1e2016db18c55ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.443658] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.468270] env[61964]: DEBUG nova.network.neutron [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.551749] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.551924] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.552092] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.591891] env[61964]: DEBUG nova.compute.manager [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Received event network-changed-3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 758.592011] env[61964]: DEBUG nova.compute.manager [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Refreshing instance network info cache due to event network-changed-3038d60e-1ce8-479d-ac96-8b597d6d8d00. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 758.592231] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] Acquiring lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.592379] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] Acquired lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.592546] env[61964]: DEBUG nova.network.neutron [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Refreshing network info cache for port 3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 758.616364] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040828, 'name': ReconfigVM_Task, 'duration_secs': 0.926195} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.616758] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Reconfigured VM instance instance-00000035 to attach disk [datastore1] fb538d67-47c3-4639-a642-b3358f97df31/fb538d67-47c3-4639-a642-b3358f97df31.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.617602] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2b7b218-581d-4dc6-99bf-25c830a7b27a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.635515] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 758.635515] env[61964]: value = "task-1040829" [ 758.635515] env[61964]: _type = "Task" [ 758.635515] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.645346] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040829, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.768814] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Successfully created port: 4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.860636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c168dc4a-fe7c-4f66-a7c2-b5c8db0607aa tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.864847] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 23.954s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.864847] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 758.971018] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f6d9701-d9ad-47a2-b751-d6500214f552 req-b601be4f-e2c9-478c-956d-b60ce66a4c94 service nova] Releasing lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.971130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.971346] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.093763] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.141222] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040829, 'name': Rename_Task, 'duration_secs': 0.149698} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.141754] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.141827] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61480c2a-eb69-458e-87e5-d80798b86df0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.153053] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 759.153053] env[61964]: value = "task-1040830" [ 759.153053] env[61964]: _type = "Task" [ 759.153053] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.161678] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.314301] env[61964]: DEBUG nova.network.neutron [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updating instance_info_cache with network_info: [{"id": "1e892208-7628-4004-8011-abf5a4a211fc", "address": "fa:16:3e:71:ec:23", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e892208-76", "ovs_interfaceid": "1e892208-7628-4004-8011-abf5a4a211fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.368036] env[61964]: DEBUG nova.objects.instance [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lazy-loading 'migration_context' on Instance uuid 65566c20-6d69-471c-b098-3c30c01d9955 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 759.402947] env[61964]: DEBUG nova.network.neutron [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updated VIF entry in instance network info cache for port 3038d60e-1ce8-479d-ac96-8b597d6d8d00. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 759.404147] env[61964]: DEBUG nova.network.neutron [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updating instance_info_cache with network_info: [{"id": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "address": "fa:16:3e:60:92:36", "network": {"id": "cf9d27f7-1266-4cd4-8407-05696aa41712", "bridge": "br-int", "label": "tempest-ServersTestJSON-981760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c44820ae8513473cbbdfddb36877a9b3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "41f66e20-fd86-4158-bbdc-7a150e85e844", "external-id": "nsx-vlan-transportzone-182", "segmentation_id": 182, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3038d60e-1c", "ovs_interfaceid": "3038d60e-1ce8-479d-ac96-8b597d6d8d00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.507868] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.665638] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040830, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.785060] env[61964]: DEBUG nova.network.neutron [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Updating instance_info_cache with network_info: [{"id": "df5c5bf6-0c91-49b4-b091-0fa57f37ec46", "address": "fa:16:3e:13:4a:08", "network": {"id": "a26a0702-9080-4ffd-b5b0-7bcb18a82ce7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1359963269", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.69", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912fe686053f4633b1099dce01d72365", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf5c5bf6-0c", "ovs_interfaceid": "df5c5bf6-0c91-49b4-b091-0fa57f37ec46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e1917c9-a522-4ee0-9a2c-7808a7918985", "address": "fa:16:3e:32:67:14", "network": {"id": "7a6de089-fadc-45fa-a1fe-ed86c3964b3e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-157217924", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "912fe686053f4633b1099dce01d72365", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e1917c9-a5", "ovs_interfaceid": "3e1917c9-a522-4ee0-9a2c-7808a7918985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.817872] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.818225] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Instance network_info: |[{"id": "1e892208-7628-4004-8011-abf5a4a211fc", "address": "fa:16:3e:71:ec:23", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e892208-76", "ovs_interfaceid": "1e892208-7628-4004-8011-abf5a4a211fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 759.819901] env[61964]: DEBUG nova.compute.manager [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received event network-vif-plugged-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 759.819901] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Acquiring lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.820143] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.820350] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.820499] env[61964]: DEBUG nova.compute.manager [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] No waiting events found dispatching network-vif-plugged-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 759.820668] env[61964]: WARNING nova.compute.manager [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received unexpected event network-vif-plugged-1e892208-7628-4004-8011-abf5a4a211fc for instance with vm_state building and task_state spawning. [ 759.821018] env[61964]: DEBUG nova.compute.manager [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received event network-changed-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 759.821228] env[61964]: DEBUG nova.compute.manager [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing instance network info cache due to event network-changed-1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 759.821436] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Acquiring lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.821577] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Acquired lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.821737] env[61964]: DEBUG nova.network.neutron [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.822892] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:ec:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e892208-7628-4004-8011-abf5a4a211fc', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.830989] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Creating folder: Project (19bac13e505143b4ae5884c24c748b1a). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.832127] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18965b2e-1372-4fbd-9d5c-eb498b95f605 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.844328] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Created folder: Project (19bac13e505143b4ae5884c24c748b1a) in parent group-v230360. [ 759.844518] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Creating folder: Instances. Parent ref: group-v230451. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.844751] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-faa7517e-2a05-44fa-a05c-0c2e9e0ad43e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.853597] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Created folder: Instances in parent group-v230451. [ 759.854677] env[61964]: DEBUG oslo.service.loopingcall [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.854677] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.854677] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38704ba4-fc11-4d5e-b37e-560ae85991f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.877896] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 759.879722] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.879722] env[61964]: value = "task-1040833" [ 759.879722] env[61964]: _type = "Task" [ 759.879722] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.889785] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040833, 'name': CreateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.905220] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 759.905501] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.905704] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 759.905904] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.906075] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 759.906235] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 759.906522] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 759.906738] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 759.906923] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 759.907115] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 759.907300] env[61964]: DEBUG nova.virt.hardware [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 759.908171] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a125ad3-01de-439a-a1bb-83edb13569b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.913441] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d573f-4b4c-47c0-8d59-6418836d60f0 req-c6e1ef9f-2aee-4e00-943b-1db4b496dcb5 service nova] Releasing lock "refresh_cache-58f8d368-fa0d-4c8d-9009-e6e1c5940032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.919264] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a040b44-c690-48c3-8ddb-08046ede1149 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.165716] env[61964]: DEBUG oslo_vmware.api [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040830, 'name': PowerOnVM_Task, 'duration_secs': 0.906583} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.166021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.166266] env[61964]: INFO nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Took 8.63 seconds to spawn the instance on the hypervisor. [ 760.166455] env[61964]: DEBUG nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 760.169633] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aab8095-57fe-414e-9a09-fc17c6c7691a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.288514] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Releasing lock "refresh_cache-0156727e-60dc-4e24-99e6-96b3e735aa0e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.288514] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance network_info: |[{"id": "df5c5bf6-0c91-49b4-b091-0fa57f37ec46", "address": "fa:16:3e:13:4a:08", "network": {"id": "a26a0702-9080-4ffd-b5b0-7bcb18a82ce7", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1359963269", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.69", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912fe686053f4633b1099dce01d72365", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf5c5bf6-0c", "ovs_interfaceid": "df5c5bf6-0c91-49b4-b091-0fa57f37ec46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3e1917c9-a522-4ee0-9a2c-7808a7918985", "address": "fa:16:3e:32:67:14", "network": {"id": "7a6de089-fadc-45fa-a1fe-ed86c3964b3e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-157217924", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "912fe686053f4633b1099dce01d72365", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e1917c9-a5", "ovs_interfaceid": "3e1917c9-a522-4ee0-9a2c-7808a7918985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 760.288514] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:4a:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df5c5bf6-0c91-49b4-b091-0fa57f37ec46', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:67:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e1917c9-a522-4ee0-9a2c-7808a7918985', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.299634] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Creating folder: Project (912fe686053f4633b1099dce01d72365). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 760.299695] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1188709f-34db-4703-a8cb-3c26c8c2b050 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.302120] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c582bf5-17aa-43e6-b982-1ac7d089e0c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.309657] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682221eb-a92d-4fe5-8b58-0c034e516775 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.314649] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Created folder: Project (912fe686053f4633b1099dce01d72365) in parent group-v230360. [ 760.314912] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Creating folder: Instances. Parent ref: group-v230454. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 760.315380] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddd5b5e9-33d2-498f-afce-e89c014dfba3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.348411] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae09a7e-e54b-458e-aa68-4b84b0c6f0bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.351315] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Created folder: Instances in parent group-v230454. [ 760.351435] env[61964]: DEBUG oslo.service.loopingcall [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.352022] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 760.352444] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89dc1491-6001-49cc-bafb-9017bcd9031c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.374451] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c67e2e-fb63-47d4-9cd6-66dfa0ebbee2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.379717] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.379717] env[61964]: value = "task-1040836" [ 760.379717] env[61964]: _type = "Task" [ 760.379717] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.392065] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Successfully updated port: 4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 760.396531] env[61964]: DEBUG nova.compute.provider_tree [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.407699] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040836, 'name': CreateVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.412375] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040833, 'name': CreateVM_Task, 'duration_secs': 0.439537} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.412521] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 760.413179] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.413357] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.413705] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 760.414038] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e71347ee-a952-4006-9a3f-fd626ecc3884 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.418935] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 760.418935] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4b83e-d1e4-3e31-0fe0-ae2eb06d6c3a" [ 760.418935] env[61964]: _type = "Task" [ 760.418935] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.427901] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4b83e-d1e4-3e31-0fe0-ae2eb06d6c3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.595363] env[61964]: DEBUG nova.network.neutron [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updated VIF entry in instance network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.595957] env[61964]: DEBUG nova.network.neutron [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updating instance_info_cache with network_info: [{"id": "1e892208-7628-4004-8011-abf5a4a211fc", "address": "fa:16:3e:71:ec:23", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e892208-76", "ovs_interfaceid": "1e892208-7628-4004-8011-abf5a4a211fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.688794] env[61964]: INFO nova.compute.manager [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Took 37.49 seconds to build instance. [ 760.890589] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040836, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.901065] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.901065] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.901065] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.903112] env[61964]: DEBUG nova.scheduler.client.report [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 760.930031] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4b83e-d1e4-3e31-0fe0-ae2eb06d6c3a, 'name': SearchDatastore_Task, 'duration_secs': 0.015167} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.930150] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.931031] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.931031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.931031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.931031] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.931249] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bd8ccb4-ceb8-450b-a7a5-3137fd1c47bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.939946] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.940154] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 760.941024] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a2aa9ac-c4a2-44df-9a3a-15cddfb2a5a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.946339] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 760.946339] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522b10bc-50c9-65bf-7879-899b3b69afd8" [ 760.946339] env[61964]: _type = "Task" [ 760.946339] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.953968] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522b10bc-50c9-65bf-7879-899b3b69afd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.098965] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa9f135-1726-43eb-80bb-d419a606d788 req-b073aea6-c593-41bd-9707-f0a8f4e7fe4b service nova] Releasing lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.192098] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a6e27408-b2d3-4010-b4de-684d13d32dbf tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.434s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.394904] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040836, 'name': CreateVM_Task, 'duration_secs': 0.598558} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.395224] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 761.396399] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.396715] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.397099] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.397374] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96f06630-a935-4ba9-a60e-fd6181ba38c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.402329] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 761.402329] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e35e82-6145-7c5e-2258-83196ec878d6" [ 761.402329] env[61964]: _type = "Task" [ 761.402329] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.417910] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e35e82-6145-7c5e-2258-83196ec878d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009247} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.418062] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.418256] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.418475] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.439751] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.456802] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522b10bc-50c9-65bf-7879-899b3b69afd8, 'name': SearchDatastore_Task, 'duration_secs': 0.00874} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.459461] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27a23483-80e4-4ef6-a81e-6bfb86333e9b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.464495] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 761.464495] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d48220-3884-a91a-72d3-934b6b2f842f" [ 761.464495] env[61964]: _type = "Task" [ 761.464495] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.471984] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d48220-3884-a91a-72d3-934b6b2f842f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.638701] env[61964]: DEBUG nova.network.neutron [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Updating instance_info_cache with network_info: [{"id": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "address": "fa:16:3e:b9:fd:22", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7e78dc-a5", "ovs_interfaceid": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.694943] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 761.747793] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "fb538d67-47c3-4639-a642-b3358f97df31" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.748136] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.748357] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "fb538d67-47c3-4639-a642-b3358f97df31-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.748547] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.748723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.750799] env[61964]: INFO nova.compute.manager [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Terminating instance [ 761.843700] env[61964]: DEBUG nova.compute.manager [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Received event network-vif-plugged-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 761.843700] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Acquiring lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.843700] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.843811] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.843990] env[61964]: DEBUG nova.compute.manager [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] No waiting events found dispatching network-vif-plugged-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 761.844160] env[61964]: WARNING nova.compute.manager [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Received unexpected event network-vif-plugged-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f for instance with vm_state building and task_state spawning. [ 761.844325] env[61964]: DEBUG nova.compute.manager [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Received event network-changed-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 761.844480] env[61964]: DEBUG nova.compute.manager [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Refreshing instance network info cache due to event network-changed-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 761.844647] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Acquiring lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.915399] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.053s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.920957] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.646s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.921226] env[61964]: DEBUG nova.objects.instance [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'resources' on Instance uuid 0ea19f52-b700-405f-8c3c-fbaa3e17b36a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 761.975755] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d48220-3884-a91a-72d3-934b6b2f842f, 'name': SearchDatastore_Task, 'duration_secs': 0.009514} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.976936] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.976936] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4/6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 761.976936] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.976936] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.977242] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2792f8af-9d24-47d1-95c1-92c67b7ec690 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.979221] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c51d2bb3-5c90-44e9-89f1-cb8afee0f922 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.986132] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 761.986132] env[61964]: value = "task-1040837" [ 761.986132] env[61964]: _type = "Task" [ 761.986132] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.989836] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.990018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 761.990995] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d45ac5e2-8bb7-4692-bf12-b480e85e6592 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.996131] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 761.996131] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521040bd-caa0-3d4e-a2ab-5921bf203ac6" [ 761.996131] env[61964]: _type = "Task" [ 761.996131] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.999232] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.006854] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521040bd-caa0-3d4e-a2ab-5921bf203ac6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.142581] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.142942] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance network_info: |[{"id": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "address": "fa:16:3e:b9:fd:22", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7e78dc-a5", "ovs_interfaceid": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 762.143297] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Acquired lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.143484] env[61964]: DEBUG nova.network.neutron [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Refreshing network info cache for port 4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 762.145064] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:fd:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e7e78dc-a5cc-4704-aefd-8382de0f2c4f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.154406] env[61964]: DEBUG oslo.service.loopingcall [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.157563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 762.158141] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a1017ae-f460-4350-a135-5e0981e4e8a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.179944] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.179944] env[61964]: value = "task-1040838" [ 762.179944] env[61964]: _type = "Task" [ 762.179944] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.190709] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040838, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.223326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.255760] env[61964]: DEBUG nova.compute.manager [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 762.256155] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 762.257358] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa175718-fcd1-4c7f-be15-e47478b05c50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.268332] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 762.269208] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7268828-ce69-4109-b3df-ecfce1af2025 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.278350] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 762.278350] env[61964]: value = "task-1040839" [ 762.278350] env[61964]: _type = "Task" [ 762.278350] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.292010] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.501712] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040837, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.515086] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521040bd-caa0-3d4e-a2ab-5921bf203ac6, 'name': SearchDatastore_Task, 'duration_secs': 0.009986} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.516243] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-575a5d1c-9e09-4bc8-83ed-ef17bf3cf029 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.522354] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 762.522354] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521718e3-b545-df7c-0b3a-ddd56b352680" [ 762.522354] env[61964]: _type = "Task" [ 762.522354] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.530121] env[61964]: DEBUG nova.network.neutron [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Updated VIF entry in instance network info cache for port 4e7e78dc-a5cc-4704-aefd-8382de0f2c4f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 762.530475] env[61964]: DEBUG nova.network.neutron [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Updating instance_info_cache with network_info: [{"id": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "address": "fa:16:3e:b9:fd:22", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e7e78dc-a5", "ovs_interfaceid": "4e7e78dc-a5cc-4704-aefd-8382de0f2c4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.537292] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521718e3-b545-df7c-0b3a-ddd56b352680, 'name': SearchDatastore_Task, 'duration_secs': 0.009036} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.538348] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.538348] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0156727e-60dc-4e24-99e6-96b3e735aa0e/0156727e-60dc-4e24-99e6-96b3e735aa0e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 762.538629] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-527d41ee-fa1f-495a-98d8-f52466bb20d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.547902] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 762.547902] env[61964]: value = "task-1040840" [ 762.547902] env[61964]: _type = "Task" [ 762.547902] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.558928] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.691706] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040838, 'name': CreateVM_Task, 'duration_secs': 0.409581} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.691879] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 762.692578] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.692748] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.693087] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 762.693386] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d491cbb-abca-4a8a-b303-650a379caf85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.699839] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 762.699839] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524870eb-b5eb-2d9e-8fef-f3d2e9b75af5" [ 762.699839] env[61964]: _type = "Task" [ 762.699839] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.707443] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524870eb-b5eb-2d9e-8fef-f3d2e9b75af5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.793781] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040839, 'name': PowerOffVM_Task, 'duration_secs': 0.261671} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.794091] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 762.794276] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 762.794546] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b70a384e-281b-4163-a3ee-9cb58f997808 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.868479] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56ab660-5087-4da5-bc47-c7505c35c1a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.875766] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 762.875998] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 762.876282] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Deleting the datastore file [datastore1] fb538d67-47c3-4639-a642-b3358f97df31 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.879213] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e0e4320-983c-42d5-8a07-9ae5d42db84b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.883947] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9e7144-632d-484a-aa38-ddb3e53a5e40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.919818] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d545c5c6-a3c1-4f5f-a9d8-68397c8e576e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.922997] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for the task: (returnval){ [ 762.922997] env[61964]: value = "task-1040842" [ 762.922997] env[61964]: _type = "Task" [ 762.922997] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.931115] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf36bc9-9502-4315-96ec-4df1ea6c7c0e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.941426] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.958191] env[61964]: DEBUG nova.compute.provider_tree [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.997172] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524208} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.997406] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4/6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 762.997718] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.997906] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0b46055-f9ec-4d84-889f-466e0bee8435 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.004439] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 763.004439] env[61964]: value = "task-1040843" [ 763.004439] env[61964]: _type = "Task" [ 763.004439] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.014849] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.033097] env[61964]: DEBUG oslo_concurrency.lockutils [req-7c174860-8107-4cd5-aa74-042c15af2fbb req-2919081b-8639-49ac-92f4-09036996e9a1 service nova] Releasing lock "refresh_cache-9adedaa2-8594-4ddd-8f85-a41174207ef8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.058143] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472957} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.058360] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0156727e-60dc-4e24-99e6-96b3e735aa0e/0156727e-60dc-4e24-99e6-96b3e735aa0e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 763.058758] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.059220] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74195027-b3b4-4d8c-b364-5b5bb5d51801 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.065562] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 763.065562] env[61964]: value = "task-1040844" [ 763.065562] env[61964]: _type = "Task" [ 763.065562] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.073600] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.211960] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524870eb-b5eb-2d9e-8fef-f3d2e9b75af5, 'name': SearchDatastore_Task, 'duration_secs': 0.008218} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.212309] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.212546] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.212787] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.214474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.214474] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.214474] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6f55e1-f1fa-458f-87c9-ba98d69f29bf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.221382] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.221605] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 763.222389] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9989af4-d2f8-4669-bfb9-c4dc8291469c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.227934] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 763.227934] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b6267f-ae78-6a27-53e6-c1904023c7c1" [ 763.227934] env[61964]: _type = "Task" [ 763.227934] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.239351] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b6267f-ae78-6a27-53e6-c1904023c7c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.432846] env[61964]: DEBUG oslo_vmware.api [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Task: {'id': task-1040842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22474} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.433204] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.433422] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 763.433609] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 763.433798] env[61964]: INFO nova.compute.manager [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Took 1.18 seconds to destroy the instance on the hypervisor. [ 763.434065] env[61964]: DEBUG oslo.service.loopingcall [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.434273] env[61964]: DEBUG nova.compute.manager [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.434366] env[61964]: DEBUG nova.network.neutron [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.464438] env[61964]: DEBUG nova.scheduler.client.report [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 763.479451] env[61964]: INFO nova.compute.manager [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Swapping old allocation on dict_keys(['57b292ab-02d9-4aab-ba83-292890345a17']) held by migration 0bbcc5dd-4fc7-4431-895f-e2476f66bace for instance [ 763.503526] env[61964]: DEBUG nova.scheduler.client.report [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Overwriting current allocation {'allocations': {'57b292ab-02d9-4aab-ba83-292890345a17': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 76}}, 'project_id': '833e91665288466a9ac2141d516ab922', 'user_id': 'cb7cc75be9a44e2187090e490b8056e8', 'consumer_generation': 1} on consumer 65566c20-6d69-471c-b098-3c30c01d9955 {{(pid=61964) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 763.514399] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072641} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.514651] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 763.515395] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3747c794-51c4-4f3b-90bf-088c8b40bf30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.537172] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4/6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 763.537480] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67395aaa-e63b-45c8-83f5-fdbee15d77ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.560382] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 763.560382] env[61964]: value = "task-1040845" [ 763.560382] env[61964]: _type = "Task" [ 763.560382] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.572461] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040845, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.577066] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070297} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.580020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 763.580020] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5181a62-b391-4a5b-890e-d54ad019ed9b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.602975] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 0156727e-60dc-4e24-99e6-96b3e735aa0e/0156727e-60dc-4e24-99e6-96b3e735aa0e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 763.602975] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e50b5ddc-de30-410b-a081-8c61c59379f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.618071] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.618354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquired lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.618730] env[61964]: DEBUG nova.network.neutron [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.627231] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 763.627231] env[61964]: value = "task-1040846" [ 763.627231] env[61964]: _type = "Task" [ 763.627231] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.635142] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.739122] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b6267f-ae78-6a27-53e6-c1904023c7c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008905} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.739866] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f30afe-a7ee-4184-8632-de672f4f42b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.745282] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 763.745282] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d22eec-2777-317b-1891-e44a8d0abbfb" [ 763.745282] env[61964]: _type = "Task" [ 763.745282] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.755470] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d22eec-2777-317b-1891-e44a8d0abbfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.827214] env[61964]: DEBUG nova.compute.manager [req-dd10802e-05e6-4385-aecf-17f2fbf5643a req-c0b4bebc-e8ce-423b-88cd-5c9bcca25896 service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Received event network-vif-deleted-307bc2ba-d780-4df4-a420-4aa33275a001 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 763.827214] env[61964]: INFO nova.compute.manager [req-dd10802e-05e6-4385-aecf-17f2fbf5643a req-c0b4bebc-e8ce-423b-88cd-5c9bcca25896 service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Neutron deleted interface 307bc2ba-d780-4df4-a420-4aa33275a001; detaching it from the instance and deleting it from the info cache [ 763.827214] env[61964]: DEBUG nova.network.neutron [req-dd10802e-05e6-4385-aecf-17f2fbf5643a req-c0b4bebc-e8ce-423b-88cd-5c9bcca25896 service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.971742] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.974089] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.581s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.975835] env[61964]: INFO nova.compute.claims [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.993882] env[61964]: INFO nova.scheduler.client.report [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance 0ea19f52-b700-405f-8c3c-fbaa3e17b36a [ 764.071774] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040845, 'name': ReconfigVM_Task, 'duration_secs': 0.47607} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.072148] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4/6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.073525] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c29874ca-c518-4e9e-8a9e-3fab6be405ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.080080] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 764.080080] env[61964]: value = "task-1040847" [ 764.080080] env[61964]: _type = "Task" [ 764.080080] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.090264] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040847, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.135173] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040846, 'name': ReconfigVM_Task, 'duration_secs': 0.30208} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.135499] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 0156727e-60dc-4e24-99e6-96b3e735aa0e/0156727e-60dc-4e24-99e6-96b3e735aa0e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.136201] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45e3ff2f-6fce-4d67-b745-76f29be87c10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.142148] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 764.142148] env[61964]: value = "task-1040848" [ 764.142148] env[61964]: _type = "Task" [ 764.142148] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.149827] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040848, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.254831] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d22eec-2777-317b-1891-e44a8d0abbfb, 'name': SearchDatastore_Task, 'duration_secs': 0.013587} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.255156] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.255477] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 9adedaa2-8594-4ddd-8f85-a41174207ef8/9adedaa2-8594-4ddd-8f85-a41174207ef8.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 764.255742] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-008e78cb-1943-46f2-a77c-1a9a5e77f70c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.262514] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 764.262514] env[61964]: value = "task-1040849" [ 764.262514] env[61964]: _type = "Task" [ 764.262514] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.271285] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040849, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.304293] env[61964]: DEBUG nova.network.neutron [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.329800] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe4319b0-7446-4c65-93a0-e587af4a9710 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.341042] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889a39f1-99ff-4f6d-9ed5-7da309af32d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.359494] env[61964]: DEBUG nova.network.neutron [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [{"id": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "address": "fa:16:3e:5c:a6:73", "network": {"id": "5a666648-d68a-4890-807f-401748218022", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.55", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "c5db75cf61c441d396a6af209d5f2d11", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd3c6b64-aba2-4bdc-a693-3b4dff3ed861", "external-id": "nsx-vlan-transportzone-600", "segmentation_id": 600, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03f523b6-4f", "ovs_interfaceid": "03f523b6-4fc2-4fe2-a751-d33c00f6849f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.375183] env[61964]: DEBUG nova.compute.manager [req-dd10802e-05e6-4385-aecf-17f2fbf5643a req-c0b4bebc-e8ce-423b-88cd-5c9bcca25896 service nova] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Detach interface failed, port_id=307bc2ba-d780-4df4-a420-4aa33275a001, reason: Instance fb538d67-47c3-4639-a642-b3358f97df31 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 764.375999] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Releasing lock "refresh_cache-65566c20-6d69-471c-b098-3c30c01d9955" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.377629] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 764.377629] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb8aa4ab-93e6-4edd-8be7-762c8cf5e45a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.383756] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 764.383756] env[61964]: value = "task-1040850" [ 764.383756] env[61964]: _type = "Task" [ 764.383756] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.393152] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.506381] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aa248f70-b8a9-4935-bc27-2f68cdab6828 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "0ea19f52-b700-405f-8c3c-fbaa3e17b36a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.950s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.593930] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040847, 'name': Rename_Task, 'duration_secs': 0.135657} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.594269] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 764.594599] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0011673b-93f5-44e2-84f5-b037cf14cb67 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.602777] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 764.602777] env[61964]: value = "task-1040851" [ 764.602777] env[61964]: _type = "Task" [ 764.602777] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.614159] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.652572] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040848, 'name': Rename_Task, 'duration_secs': 0.148801} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.653540] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 764.653848] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad144c2d-773b-41c5-b3e1-811bdefd83fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.662020] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 764.662020] env[61964]: value = "task-1040852" [ 764.662020] env[61964]: _type = "Task" [ 764.662020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.671936] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.773771] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040849, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494246} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.773771] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 9adedaa2-8594-4ddd-8f85-a41174207ef8/9adedaa2-8594-4ddd-8f85-a41174207ef8.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 764.773771] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.773771] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0761686-1b22-45d6-9a74-ade7514526fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.779787] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 764.779787] env[61964]: value = "task-1040853" [ 764.779787] env[61964]: _type = "Task" [ 764.779787] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.789441] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040853, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.807142] env[61964]: INFO nova.compute.manager [-] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Took 1.37 seconds to deallocate network for instance. [ 764.895691] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040850, 'name': PowerOffVM_Task, 'duration_secs': 0.276116} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.896112] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 764.897030] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:21:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='4d7c58f5-7642-43b5-a3cb-6aff98a75b9e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-994792262',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 764.897216] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.897393] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 764.897584] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.897738] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 764.897892] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 764.898117] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 764.898311] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 764.898486] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 764.898650] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 764.898823] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 764.904669] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-177552dc-175d-4c9d-bc3d-a8350f9fdc42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.919896] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 764.919896] env[61964]: value = "task-1040854" [ 764.919896] env[61964]: _type = "Task" [ 764.919896] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.928057] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.114022] env[61964]: DEBUG oslo_vmware.api [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040851, 'name': PowerOnVM_Task, 'duration_secs': 0.505847} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.114401] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 765.114662] env[61964]: INFO nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Took 7.81 seconds to spawn the instance on the hypervisor. [ 765.114915] env[61964]: DEBUG nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 765.115863] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225e80c1-8788-44cd-920d-58771be7a758 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.179759] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040852, 'name': PowerOnVM_Task} progress is 96%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.290923] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040853, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059904} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.291253] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 765.292107] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20eb67f7-4162-4bb1-844e-f494f0b1bd61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.316765] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 9adedaa2-8594-4ddd-8f85-a41174207ef8/9adedaa2-8594-4ddd-8f85-a41174207ef8.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 765.321288] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.321288] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9ab827c-dd51-4a8a-8621-5d169016f42c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.340674] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 765.340674] env[61964]: value = "task-1040855" [ 765.340674] env[61964]: _type = "Task" [ 765.340674] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.349127] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.439249] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040854, 'name': ReconfigVM_Task, 'duration_secs': 0.146851} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.439563] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473c2f2e-73a7-4360-a77c-877a5df326ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.458795] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:21:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='4d7c58f5-7642-43b5-a3cb-6aff98a75b9e',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-994792262',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 765.459073] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.459955] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 765.459955] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.459955] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 765.459955] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 765.459955] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 765.460228] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 765.460228] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 765.460539] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 765.460732] env[61964]: DEBUG nova.virt.hardware [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 765.464313] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70f65c9c-7194-47cb-9ac5-7823e3379de4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.470032] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 765.470032] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5294de0a-4092-4b91-7547-7b206e07819c" [ 765.470032] env[61964]: _type = "Task" [ 765.470032] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.477306] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5294de0a-4092-4b91-7547-7b206e07819c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.502860] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7291b1-ede3-46cc-bd65-a185cb058bf5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.510467] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e06d92-19e3-425e-8b2b-168d8fda0d68 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.541678] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a286d91f-f9a2-4ae4-81be-ec0601174f6e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.549496] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff4c4aa-d620-4aac-8c39-6e739a2ebe26 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.563346] env[61964]: DEBUG nova.compute.provider_tree [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.643496] env[61964]: INFO nova.compute.manager [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Took 37.73 seconds to build instance. [ 765.673567] env[61964]: DEBUG oslo_vmware.api [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040852, 'name': PowerOnVM_Task, 'duration_secs': 0.590833} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.674033] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 765.674354] env[61964]: INFO nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Took 11.58 seconds to spawn the instance on the hypervisor. [ 765.674647] env[61964]: DEBUG nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 765.675622] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8f6688-b09a-44c2-a8f6-05080236d1fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.816346] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.816346] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.818290] env[61964]: DEBUG nova.compute.manager [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received event network-changed-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 765.818290] env[61964]: DEBUG nova.compute.manager [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing instance network info cache due to event network-changed-1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 765.819120] env[61964]: DEBUG oslo_concurrency.lockutils [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] Acquiring lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.819120] env[61964]: DEBUG oslo_concurrency.lockutils [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] Acquired lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.819120] env[61964]: DEBUG nova.network.neutron [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.851415] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.979652] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5294de0a-4092-4b91-7547-7b206e07819c, 'name': SearchDatastore_Task, 'duration_secs': 0.030549} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.985524] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfiguring VM instance instance-00000022 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 765.985524] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcd1185b-f024-43cf-bee7-741cde4b5a85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.005691] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 766.005691] env[61964]: value = "task-1040856" [ 766.005691] env[61964]: _type = "Task" [ 766.005691] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.013511] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.067148] env[61964]: DEBUG nova.scheduler.client.report [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.149786] env[61964]: DEBUG oslo_concurrency.lockutils [None req-56e4d2f6-c1ee-44ac-84b0-70cec329561e tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.976s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.195884] env[61964]: INFO nova.compute.manager [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Took 40.06 seconds to build instance. [ 766.352947] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.515424] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040856, 'name': ReconfigVM_Task, 'duration_secs': 0.254729} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.515705] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfigured VM instance instance-00000022 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 766.516666] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec236cf8-e32a-45ac-ac06-5b2a7fcef362 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.540947] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.544328] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faf1c14a-b436-42db-84ec-8a81422b556e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.563863] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 766.563863] env[61964]: value = "task-1040857" [ 766.563863] env[61964]: _type = "Task" [ 766.563863] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.572403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.573102] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 766.579558] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.456s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.579724] env[61964]: DEBUG nova.objects.instance [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'resources' on Instance uuid df62ed65-0a89-4f04-9b5a-f5f4214e6bd5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 766.581162] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040857, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.652016] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 766.667417] env[61964]: DEBUG nova.network.neutron [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updated VIF entry in instance network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.667734] env[61964]: DEBUG nova.network.neutron [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updating instance_info_cache with network_info: [{"id": "1e892208-7628-4004-8011-abf5a4a211fc", "address": "fa:16:3e:71:ec:23", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e892208-76", "ovs_interfaceid": "1e892208-7628-4004-8011-abf5a4a211fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.699146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6fd7f169-92cd-417e-94e1-9c0b99b9af73 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.664s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.819420] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.819700] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.820107] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.820174] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.820359] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.823896] env[61964]: INFO nova.compute.manager [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Terminating instance [ 766.851709] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040855, 'name': ReconfigVM_Task, 'duration_secs': 1.155866} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.852216] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 9adedaa2-8594-4ddd-8f85-a41174207ef8/9adedaa2-8594-4ddd-8f85-a41174207ef8.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.852865] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53836bde-d257-4cb8-920c-7e5c21853323 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.859596] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 766.859596] env[61964]: value = "task-1040858" [ 766.859596] env[61964]: _type = "Task" [ 766.859596] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.868554] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040858, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.075998] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040857, 'name': ReconfigVM_Task, 'duration_secs': 0.273366} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.076322] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955/65566c20-6d69-471c-b098-3c30c01d9955.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.077161] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95d7e21-d32d-436d-ab3c-a36fbebe7621 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.097603] env[61964]: DEBUG nova.compute.utils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.103517] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 767.103570] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.107157] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ff98f3-3b96-49b9-b433-3fe767d487a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.143765] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7e89ca-3a0a-4b1c-af0d-8841a3eae148 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.176242] env[61964]: DEBUG oslo_concurrency.lockutils [req-fca3bc0e-0549-421a-80a9-3da2e763ec71 req-2bb5970c-1832-468b-8871-6f1ec83afa5d service nova] Releasing lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.181874] env[61964]: DEBUG nova.policy [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.185404] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57046753-a29a-43b3-869d-31cb723b862a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.192511] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 767.192782] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50a40476-2534-4cb4-abf4-3844d0f0e2fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.198977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.203317] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 767.205847] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 767.205847] env[61964]: value = "task-1040859" [ 767.205847] env[61964]: _type = "Task" [ 767.205847] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.217044] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.357370] env[61964]: DEBUG nova.compute.manager [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 767.357370] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 767.357370] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b403ae7-a8fe-49bd-82ec-db24aece5764 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.357370] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 767.357370] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92e9b285-5212-487c-ac49-044e70d239aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.357370] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 767.357370] env[61964]: value = "task-1040860" [ 767.357370] env[61964]: _type = "Task" [ 767.357370] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.357370] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.368370] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040858, 'name': Rename_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.603999] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 767.655746] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac38a3c-3327-469a-93f8-69917aaab8ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.667106] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d54e37-8be6-4a39-bef1-9802f5687bb2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.699018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87662de4-77aa-4673-9c2a-aeee32939485 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.711062] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611e7b75-d7f1-4ce5-8784-29a676c25267 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.734087] env[61964]: DEBUG nova.compute.provider_tree [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.737146] env[61964]: DEBUG oslo_vmware.api [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040859, 'name': PowerOnVM_Task, 'duration_secs': 0.413765} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.737146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.737146] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.748911] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Successfully created port: 0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.856460] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040860, 'name': PowerOffVM_Task, 'duration_secs': 0.212021} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.856654] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 767.856854] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 767.857137] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b73de33c-928f-43d3-a723-fd744d4366be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.869023] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040858, 'name': Rename_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.959234] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 767.959234] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 767.959234] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Deleting the datastore file [datastore1] 0156727e-60dc-4e24-99e6-96b3e735aa0e {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 767.959234] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66e00dbd-11db-4411-a01f-34bf0c6b3a94 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.965827] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for the task: (returnval){ [ 767.965827] env[61964]: value = "task-1040862" [ 767.965827] env[61964]: _type = "Task" [ 767.965827] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.974302] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.022513] env[61964]: DEBUG nova.compute.manager [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received event network-changed-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 768.022889] env[61964]: DEBUG nova.compute.manager [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing instance network info cache due to event network-changed-1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 768.023050] env[61964]: DEBUG oslo_concurrency.lockutils [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] Acquiring lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.023245] env[61964]: DEBUG oslo_concurrency.lockutils [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] Acquired lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.023455] env[61964]: DEBUG nova.network.neutron [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Refreshing network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.237596] env[61964]: DEBUG nova.scheduler.client.report [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.371864] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040858, 'name': Rename_Task, 'duration_secs': 1.174087} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.371864] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 768.371864] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6726e9ef-f4aa-4167-ad3e-97700cc45b69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.378387] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 768.378387] env[61964]: value = "task-1040863" [ 768.378387] env[61964]: _type = "Task" [ 768.378387] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.385752] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.476038] env[61964]: DEBUG oslo_vmware.api [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Task: {'id': task-1040862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29902} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.476038] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 768.476181] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 768.476443] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 768.476697] env[61964]: INFO nova.compute.manager [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 768.477067] env[61964]: DEBUG oslo.service.loopingcall [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.477310] env[61964]: DEBUG nova.compute.manager [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 768.477437] env[61964]: DEBUG nova.network.neutron [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.614234] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 768.655314] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 768.655594] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.655766] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 768.656019] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.656186] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 768.656348] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 768.656563] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 768.656729] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 768.656989] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 768.657196] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 768.657425] env[61964]: DEBUG nova.virt.hardware [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 768.658410] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4401d60b-c9dd-4874-ba25-7909c2027a54 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.667617] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be04327-f3e7-48da-967b-eb86f01ad206 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.748608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.167s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.754931] env[61964]: INFO nova.compute.manager [None req-f778acfb-492a-49a0-9b9e-7ca5247b118b tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance to original state: 'active' [ 768.759542] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.377s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.759773] env[61964]: DEBUG nova.objects.instance [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lazy-loading 'resources' on Instance uuid e5e7b687-5691-4b1e-b113-2b8e83b23662 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 768.783091] env[61964]: INFO nova.scheduler.client.report [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocations for instance df62ed65-0a89-4f04-9b5a-f5f4214e6bd5 [ 768.889610] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.142283] env[61964]: DEBUG nova.network.neutron [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updated VIF entry in instance network info cache for port 1e892208-7628-4004-8011-abf5a4a211fc. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 769.143608] env[61964]: DEBUG nova.network.neutron [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updating instance_info_cache with network_info: [{"id": "1e892208-7628-4004-8011-abf5a4a211fc", "address": "fa:16:3e:71:ec:23", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e892208-76", "ovs_interfaceid": "1e892208-7628-4004-8011-abf5a4a211fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.292713] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1f9dcba-5ebd-4e21-90b3-161d0251bb0a tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "df62ed65-0a89-4f04-9b5a-f5f4214e6bd5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.864s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.392491] env[61964]: DEBUG oslo_vmware.api [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040863, 'name': PowerOnVM_Task, 'duration_secs': 0.971667} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.393152] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 769.393152] env[61964]: INFO nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Took 9.51 seconds to spawn the instance on the hypervisor. [ 769.393269] env[61964]: DEBUG nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 769.394375] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be608599-6a35-4cc1-9f94-493ab544536f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.461093] env[61964]: DEBUG nova.compute.manager [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-plugged-0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 769.461093] env[61964]: DEBUG oslo_concurrency.lockutils [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.461093] env[61964]: DEBUG oslo_concurrency.lockutils [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.461093] env[61964]: DEBUG oslo_concurrency.lockutils [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.461093] env[61964]: DEBUG nova.compute.manager [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] No waiting events found dispatching network-vif-plugged-0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 769.461093] env[61964]: WARNING nova.compute.manager [req-949b65bd-0aa8-4cd1-8082-508fcb3cd6e8 req-c85e85c1-9a1a-4eca-aff2-e35b8c784f76 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received unexpected event network-vif-plugged-0db516e7-e7df-4355-b5d0-0d2dad89e921 for instance with vm_state building and task_state spawning. [ 769.511178] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Successfully updated port: 0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.646099] env[61964]: DEBUG oslo_concurrency.lockutils [req-37694c78-212f-42a1-85e1-d03ae31c4b61 req-c986b857-dc43-4200-accb-59f2bca91b39 service nova] Releasing lock "refresh_cache-6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.705299] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef9246b-77c7-4027-8104-4cfae8e72164 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.713406] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286d9550-28e0-41ad-88f2-88c843bd7ba5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.743754] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645b5a39-8442-4d65-8df1-35cbec181865 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.751354] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb9e881-d14d-4133-b3d3-a6d2051c90b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.765676] env[61964]: DEBUG nova.compute.provider_tree [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.914607] env[61964]: DEBUG nova.network.neutron [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.918948] env[61964]: INFO nova.compute.manager [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Took 36.79 seconds to build instance. [ 770.017758] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.017921] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.018089] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.046447] env[61964]: DEBUG nova.compute.manager [req-c690dc3b-9540-4b56-8db8-d715bd7cd577 req-0e7d9d71-7049-4bcd-95de-dc13645e06f8 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-vif-deleted-df5c5bf6-0c91-49b4-b091-0fa57f37ec46 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 770.046659] env[61964]: DEBUG nova.compute.manager [req-c690dc3b-9540-4b56-8db8-d715bd7cd577 req-0e7d9d71-7049-4bcd-95de-dc13645e06f8 service nova] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Received event network-vif-deleted-3e1917c9-a522-4ee0-9a2c-7808a7918985 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 770.269077] env[61964]: DEBUG nova.scheduler.client.report [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.302803] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "65566c20-6d69-471c-b098-3c30c01d9955" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.303096] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.303313] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "65566c20-6d69-471c-b098-3c30c01d9955-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.303990] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.304202] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.306375] env[61964]: INFO nova.compute.manager [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Terminating instance [ 770.419618] env[61964]: INFO nova.compute.manager [-] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Took 1.94 seconds to deallocate network for instance. [ 770.421712] env[61964]: DEBUG oslo_concurrency.lockutils [None req-39a29827-a6d3-4bfd-8f42-0754286315a7 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.733s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.437059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "2a185e4b-c636-418a-a75e-2cf158f550c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.437360] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.551017] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.578447] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ade1563-a520-46a5-b451-19e545e070fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.587323] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Suspending the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 770.588081] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-2e65f4bb-4175-46cd-bebd-c284d8192ff0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.596039] env[61964]: DEBUG oslo_vmware.api [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 770.596039] env[61964]: value = "task-1040864" [ 770.596039] env[61964]: _type = "Task" [ 770.596039] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.607137] env[61964]: DEBUG oslo_vmware.api [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040864, 'name': SuspendVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.738722] env[61964]: DEBUG nova.network.neutron [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.774812] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.015s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.777471] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.628s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.779139] env[61964]: INFO nova.compute.claims [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.803236] env[61964]: INFO nova.scheduler.client.report [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Deleted allocations for instance e5e7b687-5691-4b1e-b113-2b8e83b23662 [ 770.811219] env[61964]: DEBUG nova.compute.manager [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 770.811572] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 770.812666] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe59e30-f694-4bcf-b5e1-a219e4905325 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.822576] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 770.822722] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-800e73ff-ecbb-455b-bf20-124f9a63f848 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.832200] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 770.832200] env[61964]: value = "task-1040865" [ 770.832200] env[61964]: _type = "Task" [ 770.832200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.844212] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.928066] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.940077] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 771.109332] env[61964]: DEBUG oslo_vmware.api [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040864, 'name': SuspendVM_Task} progress is 50%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.241434] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.241698] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Instance network_info: |[{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 771.242167] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:29:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0db516e7-e7df-4355-b5d0-0d2dad89e921', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.250139] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Creating folder: Project (9d4175848e8e492ba4f749e464b693ec). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.250501] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-977e666c-754d-4c93-a8ed-29c846cc7ec4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.263367] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Created folder: Project (9d4175848e8e492ba4f749e464b693ec) in parent group-v230360. [ 771.264048] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Creating folder: Instances. Parent ref: group-v230458. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.264048] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0c9cbe7-6e0d-4d6a-8094-7134e102a740 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.276141] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Created folder: Instances in parent group-v230458. [ 771.276452] env[61964]: DEBUG oslo.service.loopingcall [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.276682] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 771.276911] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5feed35-47da-476a-8f57-f66bdb728a75 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.303333] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.303333] env[61964]: value = "task-1040868" [ 771.303333] env[61964]: _type = "Task" [ 771.303333] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.314993] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040868, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.315607] env[61964]: DEBUG oslo_concurrency.lockutils [None req-449bfdc0-d580-4c14-b406-49ff1c229178 tempest-ServersNegativeTestMultiTenantJSON-384094890 tempest-ServersNegativeTestMultiTenantJSON-384094890-project-member] Lock "e5e7b687-5691-4b1e-b113-2b8e83b23662" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.376s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.345415] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040865, 'name': PowerOffVM_Task, 'duration_secs': 0.252567} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.345709] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 771.345889] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 771.346604] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2833a8c-687a-46d3-92bb-93dd00395f91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.463953] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.484597] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 771.484928] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 771.485179] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Deleting the datastore file [datastore2] 65566c20-6d69-471c-b098-3c30c01d9955 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.485538] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c3a700b-713e-4f9f-a01f-048a6a4d6bef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.493740] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for the task: (returnval){ [ 771.493740] env[61964]: value = "task-1040870" [ 771.493740] env[61964]: _type = "Task" [ 771.493740] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.502786] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.529550] env[61964]: DEBUG nova.compute.manager [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-changed-0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 771.529764] env[61964]: DEBUG nova.compute.manager [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing instance network info cache due to event network-changed-0db516e7-e7df-4355-b5d0-0d2dad89e921. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 771.529982] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.530177] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.530339] env[61964]: DEBUG nova.network.neutron [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing network info cache for port 0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.608637] env[61964]: DEBUG oslo_vmware.api [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040864, 'name': SuspendVM_Task} progress is 50%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.812243] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040868, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.010906] env[61964]: DEBUG oslo_vmware.api [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Task: {'id': task-1040870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.410683} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.011277] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.011467] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 772.011648] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 772.011826] env[61964]: INFO nova.compute.manager [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Took 1.20 seconds to destroy the instance on the hypervisor. [ 772.012101] env[61964]: DEBUG oslo.service.loopingcall [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.012309] env[61964]: DEBUG nova.compute.manager [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 772.012401] env[61964]: DEBUG nova.network.neutron [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.122346] env[61964]: DEBUG oslo_vmware.api [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040864, 'name': SuspendVM_Task, 'duration_secs': 1.174637} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.122772] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Suspended the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 772.122845] env[61964]: DEBUG nova.compute.manager [None req-695b501a-c8a8-4c07-9764-46e386ebf084 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 772.123760] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f491dd8-d323-43ce-be6f-04607ede1419 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.318485] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040868, 'name': CreateVM_Task, 'duration_secs': 0.597417} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.321843] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 772.323454] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.323754] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.324254] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.324635] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-544683fc-4960-4259-81c9-f2624a2d7da3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.330183] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 772.330183] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bd4192-82a9-0a90-49b9-8df56aef31ac" [ 772.330183] env[61964]: _type = "Task" [ 772.330183] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.341480] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bd4192-82a9-0a90-49b9-8df56aef31ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.345917] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884c5ed2-7503-4be5-8903-ab9cb42f5655 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.348616] env[61964]: DEBUG nova.network.neutron [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updated VIF entry in instance network info cache for port 0db516e7-e7df-4355-b5d0-0d2dad89e921. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.349242] env[61964]: DEBUG nova.network.neutron [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.355960] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5c01e1-bc78-48c9-88e4-687bb9632241 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.390829] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f963f889-e1cb-40c8-a56c-7ada831df29d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.401777] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14a1dc9-677e-48f7-bfa2-9b42d6b11b79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.419347] env[61964]: DEBUG nova.compute.provider_tree [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.844715] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bd4192-82a9-0a90-49b9-8df56aef31ac, 'name': SearchDatastore_Task, 'duration_secs': 0.010297} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.846030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.846030] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.846030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.846030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.847035] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.847035] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b14eaf08-a69c-443f-a947-e6a75784608f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.851448] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0946642-2808-450e-b5ae-2167bc3a06ae req-af89cc3f-a921-49d4-8614-88ad76586f79 service nova] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.860605] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.860972] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 772.861889] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4148bbf3-1e79-464b-9578-fae9b297c10a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.868605] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 772.868605] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522acf4a-a297-a294-5a8b-090115e13572" [ 772.868605] env[61964]: _type = "Task" [ 772.868605] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.879371] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522acf4a-a297-a294-5a8b-090115e13572, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.926916] env[61964]: DEBUG nova.scheduler.client.report [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 772.940140] env[61964]: DEBUG nova.network.neutron [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.383676] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522acf4a-a297-a294-5a8b-090115e13572, 'name': SearchDatastore_Task, 'duration_secs': 0.038407} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.383676] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2466b05-fb66-4bb5-90e7-317f2dce7a8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.390331] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 773.390331] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525abb35-7d8b-3af1-a6f9-30c81e0fa61c" [ 773.390331] env[61964]: _type = "Task" [ 773.390331] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.399958] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525abb35-7d8b-3af1-a6f9-30c81e0fa61c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.434376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.657s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.435198] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 773.439376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.580s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.439476] env[61964]: DEBUG nova.objects.instance [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lazy-loading 'resources' on Instance uuid b50e683f-79b2-419a-ac95-94908adb2b3e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.444100] env[61964]: INFO nova.compute.manager [-] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Took 1.43 seconds to deallocate network for instance. [ 773.563149] env[61964]: DEBUG nova.compute.manager [req-94458360-1dc1-4d8d-9191-ae554e57244c req-b0137ba0-8ffd-4857-8a80-6506658c32ac service nova] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Received event network-vif-deleted-03f523b6-4fc2-4fe2-a751-d33c00f6849f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 773.904239] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525abb35-7d8b-3af1-a6f9-30c81e0fa61c, 'name': SearchDatastore_Task, 'duration_secs': 0.023728} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.904742] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.905150] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] e93f98b3-4763-4f02-abd3-c24a9ab8dc17/e93f98b3-4763-4f02-abd3-c24a9ab8dc17.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 773.905599] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-940663db-848e-4e56-9ec8-fad34070f0bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.917024] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 773.917024] env[61964]: value = "task-1040871" [ 773.917024] env[61964]: _type = "Task" [ 773.917024] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.931774] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.946079] env[61964]: DEBUG nova.compute.utils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.949886] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 773.950158] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.959028] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.007728] env[61964]: DEBUG nova.policy [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '599501bcaf874b45a3faa4b4ccfb11e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b772e2152b674da8bdc2cb4f726f6772', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.073304] env[61964]: DEBUG nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 774.074393] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feda981c-930b-4348-8699-1fe99422ebc6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.428208] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26500f6e-c49a-4a2e-859e-046b79829484 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.436120] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040871, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.443014] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa724bc0-040d-4009-9061-91ad98e26e95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.483132] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 774.487793] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87151b0-856d-478b-8cc4-3c51c5f7a685 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.499231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134faf7e-debc-4485-bf50-52a2cdae46a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.519179] env[61964]: DEBUG nova.compute.provider_tree [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.539406] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Successfully created port: 63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.592861] env[61964]: INFO nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] instance snapshotting [ 774.593106] env[61964]: WARNING nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 774.596853] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b6724e-9231-4b64-81f1-34f789c06f88 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.617997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55faa77b-6f03-4f75-be44-67c7c5e6d907 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.931753] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.668016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.932047] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] e93f98b3-4763-4f02-abd3-c24a9ab8dc17/e93f98b3-4763-4f02-abd3-c24a9ab8dc17.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.932280] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.932539] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd04afe8-1b1e-4394-bfa4-b0e1cda81a9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.940610] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 774.940610] env[61964]: value = "task-1040872" [ 774.940610] env[61964]: _type = "Task" [ 774.940610] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.949113] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040872, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.023786] env[61964]: DEBUG nova.scheduler.client.report [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 775.131433] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 775.131674] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c1877426-9c8f-4b80-93ed-fcd00e35f645 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.142767] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 775.142767] env[61964]: value = "task-1040873" [ 775.142767] env[61964]: _type = "Task" [ 775.142767] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.153652] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040873, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.450996] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040872, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.383414} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.451295] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.452140] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bed0eb8-031d-498a-af7e-22364c3d90ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.476421] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] e93f98b3-4763-4f02-abd3-c24a9ab8dc17/e93f98b3-4763-4f02-abd3-c24a9ab8dc17.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.476421] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b93b815-a609-47ad-975a-f7376af7e079 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.499186] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 775.506066] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 775.506066] env[61964]: value = "task-1040874" [ 775.506066] env[61964]: _type = "Task" [ 775.506066] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.516911] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040874, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.530472] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 775.530851] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.531096] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 775.531310] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.531503] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 775.531684] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 775.531952] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 775.532211] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 775.532453] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 775.532682] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 775.532918] env[61964]: DEBUG nova.virt.hardware [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 775.533858] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.096s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.536819] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d0e995-d053-4fb0-b981-3795b0082499 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.539948] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.102s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.541718] env[61964]: INFO nova.compute.claims [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.551764] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25235935-b75f-4dc0-8a9f-deedc98da2f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.570399] env[61964]: INFO nova.scheduler.client.report [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Deleted allocations for instance b50e683f-79b2-419a-ac95-94908adb2b3e [ 775.653461] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040873, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.018837] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040874, 'name': ReconfigVM_Task, 'duration_secs': 0.362391} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.020460] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfigured VM instance instance-00000039 to attach disk [datastore2] e93f98b3-4763-4f02-abd3-c24a9ab8dc17/e93f98b3-4763-4f02-abd3-c24a9ab8dc17.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.021968] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d568342c-4b41-4a05-8133-4eb084a97af5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.031277] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 776.031277] env[61964]: value = "task-1040875" [ 776.031277] env[61964]: _type = "Task" [ 776.031277] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.043289] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040875, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.078730] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2e8b2ee4-a42d-4f99-9504-b4f82badf519 tempest-InstanceActionsNegativeTestJSON-1565753066 tempest-InstanceActionsNegativeTestJSON-1565753066-project-member] Lock "b50e683f-79b2-419a-ac95-94908adb2b3e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.885s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.155029] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040873, 'name': CreateSnapshot_Task, 'duration_secs': 0.574636} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.155250] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 776.157049] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c306b4-d351-4ae4-95a8-f56232a6d763 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.364229] env[61964]: DEBUG nova.compute.manager [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Received event network-vif-plugged-63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 776.364468] env[61964]: DEBUG oslo_concurrency.lockutils [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] Acquiring lock "595c94f1-9dec-454d-b301-5656287f53de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.364713] env[61964]: DEBUG oslo_concurrency.lockutils [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] Lock "595c94f1-9dec-454d-b301-5656287f53de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.365610] env[61964]: DEBUG oslo_concurrency.lockutils [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] Lock "595c94f1-9dec-454d-b301-5656287f53de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.365610] env[61964]: DEBUG nova.compute.manager [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] No waiting events found dispatching network-vif-plugged-63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 776.365610] env[61964]: WARNING nova.compute.manager [req-c99f675b-a7c4-441a-b46f-3f1ddc3c117e req-f2aa0813-9801-44c2-828a-6b8970834338 service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Received unexpected event network-vif-plugged-63bb7733-05d2-4984-bbe6-57940384a3a8 for instance with vm_state building and task_state spawning. [ 776.415833] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Successfully updated port: 63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.542705] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040875, 'name': Rename_Task, 'duration_secs': 0.162439} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.543106] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.543483] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f149511-d7a7-43c0-a431-621f488364ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.553781] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 776.553781] env[61964]: value = "task-1040876" [ 776.553781] env[61964]: _type = "Task" [ 776.553781] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.563977] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.683354] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 776.685630] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c1177214-75dc-41e0-a582-a8ed1c72076e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.701509] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 776.701509] env[61964]: value = "task-1040877" [ 776.701509] env[61964]: _type = "Task" [ 776.701509] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.712282] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040877, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.894019] env[61964]: DEBUG nova.compute.manager [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 776.895019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4616bb-f4cd-49e5-b488-85a9a316360f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.918812] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.919083] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.919293] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.069920] env[61964]: DEBUG oslo_vmware.api [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1040876, 'name': PowerOnVM_Task, 'duration_secs': 0.51329} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.070662] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.070662] env[61964]: INFO nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Took 8.46 seconds to spawn the instance on the hypervisor. [ 777.070662] env[61964]: DEBUG nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 777.071699] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5997b4b1-0e3f-4abe-80e4-c1c5d786507f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.076595] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6167d8c9-a7f0-4e07-9602-f9495660d3a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.091517] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a291f1a5-49e8-4211-81e8-f35a17e2cf45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.128192] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c35aeab-d3de-4c60-8121-88f448fad426 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.138515] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c95e960-6352-485d-8e85-a4699f359bab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.155267] env[61964]: DEBUG nova.compute.provider_tree [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.215729] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040877, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.412128] env[61964]: INFO nova.compute.manager [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] instance snapshotting [ 777.412827] env[61964]: DEBUG nova.objects.instance [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 777.478468] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.603921] env[61964]: INFO nova.compute.manager [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Took 40.23 seconds to build instance. [ 777.659467] env[61964]: DEBUG nova.scheduler.client.report [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 777.702887] env[61964]: DEBUG nova.network.neutron [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Updating instance_info_cache with network_info: [{"id": "63bb7733-05d2-4984-bbe6-57940384a3a8", "address": "fa:16:3e:6c:13:02", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63bb7733-05", "ovs_interfaceid": "63bb7733-05d2-4984-bbe6-57940384a3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.717545] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040877, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.920651] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d842b4-101b-4341-ba48-53681f3d0d24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.950499] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "6ee845c3-5f79-4704-8b7b-cd3770202647" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.950599] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.951133] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.951133] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.951133] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.955712] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e6d2ca-406e-4c69-898d-8002fb552f48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.959198] env[61964]: INFO nova.compute.manager [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Terminating instance [ 778.109828] env[61964]: DEBUG oslo_concurrency.lockutils [None req-012eed5a-e6e5-443c-9c92-8b19066e7df2 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.159s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.165191] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.165933] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 778.173918] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.189s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.174184] env[61964]: INFO nova.compute.claims [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.205711] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.205906] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Instance network_info: |[{"id": "63bb7733-05d2-4984-bbe6-57940384a3a8", "address": "fa:16:3e:6c:13:02", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63bb7733-05", "ovs_interfaceid": "63bb7733-05d2-4984-bbe6-57940384a3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 778.206884] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:13:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02092ea4-bae0-4e42-b0ab-abc365b4395a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63bb7733-05d2-4984-bbe6-57940384a3a8', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.217044] env[61964]: DEBUG oslo.service.loopingcall [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.227613] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 778.227613] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c269a62a-23e2-4bca-96a7-676d996c3b7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.250896] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040877, 'name': CloneVM_Task, 'duration_secs': 1.493541} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.252389] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Created linked-clone VM from snapshot [ 778.252637] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.252637] env[61964]: value = "task-1040878" [ 778.252637] env[61964]: _type = "Task" [ 778.252637] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.253372] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c57559-0ede-479e-86eb-9e0047bb5493 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.265971] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Uploading image 95e7d434-615c-44b6-a948-97f1713d7c9f {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 778.273221] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040878, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.305801] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 778.305801] env[61964]: value = "vm-230462" [ 778.305801] env[61964]: _type = "VirtualMachine" [ 778.305801] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 778.306204] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-032d4e83-7da9-4aac-b8fd-88014abb26b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.317177] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease: (returnval){ [ 778.317177] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cb57e-0314-e998-428b-edd5f1a8150c" [ 778.317177] env[61964]: _type = "HttpNfcLease" [ 778.317177] env[61964]: } obtained for exporting VM: (result){ [ 778.317177] env[61964]: value = "vm-230462" [ 778.317177] env[61964]: _type = "VirtualMachine" [ 778.317177] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 778.317437] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the lease: (returnval){ [ 778.317437] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cb57e-0314-e998-428b-edd5f1a8150c" [ 778.317437] env[61964]: _type = "HttpNfcLease" [ 778.317437] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 778.326410] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 778.326410] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cb57e-0314-e998-428b-edd5f1a8150c" [ 778.326410] env[61964]: _type = "HttpNfcLease" [ 778.326410] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 778.394667] env[61964]: DEBUG nova.compute.manager [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Received event network-changed-63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 778.394869] env[61964]: DEBUG nova.compute.manager [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Refreshing instance network info cache due to event network-changed-63bb7733-05d2-4984-bbe6-57940384a3a8. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 778.395157] env[61964]: DEBUG oslo_concurrency.lockutils [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] Acquiring lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.395337] env[61964]: DEBUG oslo_concurrency.lockutils [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] Acquired lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.395527] env[61964]: DEBUG nova.network.neutron [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Refreshing network info cache for port 63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.466143] env[61964]: DEBUG nova.compute.manager [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 778.466143] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 778.466143] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15d046ec-3d91-4bf1-990b-693214c9fd94 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.470153] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 778.470347] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-08449ca3-36df-4352-8ad1-f38eaad544b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.474925] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 778.474925] env[61964]: value = "task-1040880" [ 778.474925] env[61964]: _type = "Task" [ 778.474925] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.479639] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 778.479639] env[61964]: value = "task-1040881" [ 778.479639] env[61964]: _type = "Task" [ 778.479639] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.488439] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.494877] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040881, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.684514] env[61964]: DEBUG nova.compute.utils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.688930] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 778.689347] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.769345] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040878, 'name': CreateVM_Task, 'duration_secs': 0.379362} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.769345] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 778.770840] env[61964]: DEBUG nova.policy [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.775536] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.775722] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.776175] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.776843] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ee6c8c3-00fb-4f0b-b055-cdb3ac8a9f12 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.785666] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 778.785666] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abc37-2bbe-8f97-21f4-76aaba04c81b" [ 778.785666] env[61964]: _type = "Task" [ 778.785666] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.795330] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abc37-2bbe-8f97-21f4-76aaba04c81b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.826578] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 778.826578] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cb57e-0314-e998-428b-edd5f1a8150c" [ 778.826578] env[61964]: _type = "HttpNfcLease" [ 778.826578] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 778.826926] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 778.826926] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cb57e-0314-e998-428b-edd5f1a8150c" [ 778.826926] env[61964]: _type = "HttpNfcLease" [ 778.826926] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 778.827692] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74744ea-9327-4dfc-b566-59c9cba0625c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.836944] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 778.836944] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 778.896803] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.897082] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.897295] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.897552] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.897783] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.902789] env[61964]: INFO nova.compute.manager [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Terminating instance [ 778.935268] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6cf96256-4b5f-43d1-9b78-58dfbba0b41f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.986020] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040880, 'name': PowerOffVM_Task, 'duration_secs': 0.26588} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.987671] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 778.990085] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 778.990085] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230373', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'name': 'volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6ee845c3-5f79-4704-8b7b-cd3770202647', 'attached_at': '', 'detached_at': '', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'serial': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 778.990085] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de90dfed-9dab-4c4d-ae6e-f1791d440a6f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.999032] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040881, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.019141] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9858c2-393f-423f-a341-62ffbc03b0bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.027805] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6306165-4c07-4b87-b315-c4f1faeb24e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.050892] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb02346-4c09-4fd4-ac10-98ea88326505 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.067440] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] The volume has not been displaced from its original location: [datastore1] volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9/volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 779.074458] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Reconfiguring VM instance instance-0000002d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 779.074458] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-689e8ff1-0414-45d1-83c5-44e36f336f0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.098021] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 779.098021] env[61964]: value = "task-1040882" [ 779.098021] env[61964]: _type = "Task" [ 779.098021] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.104951] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040882, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.192542] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 779.239673] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Successfully created port: 0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.296810] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abc37-2bbe-8f97-21f4-76aaba04c81b, 'name': SearchDatastore_Task, 'duration_secs': 0.01338} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.297221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.297510] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.297779] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.297930] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.298140] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.298431] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42829b1d-ed8a-4851-868b-f9561bcd6c2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.310448] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.310595] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.313177] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-958dc6ba-6c17-40a8-b3eb-181f9485392b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.317817] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 779.317817] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52828b2a-fc4f-1007-ff45-289f69fab661" [ 779.317817] env[61964]: _type = "Task" [ 779.317817] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.331634] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52828b2a-fc4f-1007-ff45-289f69fab661, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.406765] env[61964]: DEBUG nova.compute.manager [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 779.408451] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.410176] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366f7259-c201-4d2c-897e-3e3753916732 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.413954] env[61964]: DEBUG nova.network.neutron [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Updated VIF entry in instance network info cache for port 63bb7733-05d2-4984-bbe6-57940384a3a8. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.416808] env[61964]: DEBUG nova.network.neutron [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Updating instance_info_cache with network_info: [{"id": "63bb7733-05d2-4984-bbe6-57940384a3a8", "address": "fa:16:3e:6c:13:02", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63bb7733-05", "ovs_interfaceid": "63bb7733-05d2-4984-bbe6-57940384a3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.428945] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.432500] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c88c897-831e-4601-9867-bc1630cd075a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.443789] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 779.443789] env[61964]: value = "task-1040883" [ 779.443789] env[61964]: _type = "Task" [ 779.443789] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.460161] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.511276] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040881, 'name': CreateSnapshot_Task, 'duration_secs': 0.830078} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.511777] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 779.512887] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c15aa4e-2239-4686-acdc-30e2d58b88b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.614848] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040882, 'name': ReconfigVM_Task, 'duration_secs': 0.217348} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.615588] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Reconfigured VM instance instance-0000002d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 779.628423] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbddb006-6336-4e56-b269-8a6d7a3e3fae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.661256] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 779.661256] env[61964]: value = "task-1040884" [ 779.661256] env[61964]: _type = "Task" [ 779.661256] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.682487] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040884, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.832138] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52828b2a-fc4f-1007-ff45-289f69fab661, 'name': SearchDatastore_Task, 'duration_secs': 0.020462} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.833316] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8677025a-9af3-42d5-b6eb-cfc155cfda7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.840185] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 779.840185] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52136db8-d481-948d-a123-c5b02583d97e" [ 779.840185] env[61964]: _type = "Task" [ 779.840185] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.853892] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52136db8-d481-948d-a123-c5b02583d97e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.921127] env[61964]: DEBUG oslo_concurrency.lockutils [req-28cab04d-fe26-4e08-b2ae-dd338c26b1ef req-1029274e-c19f-4921-963d-caea6219f06e service nova] Releasing lock "refresh_cache-595c94f1-9dec-454d-b301-5656287f53de" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.921498] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f27f906-e74b-4956-82e7-e72fae2c8ced {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.930179] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c344f6-b6d5-4ee2-b8fc-119c4742c006 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.973167] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce975e28-da22-4003-91da-4fbfeb3a8674 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.982679] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040883, 'name': PowerOffVM_Task, 'duration_secs': 0.293277} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.984877] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.985277] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.985916] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-148fcbdf-bf01-40bb-882b-e576de00d82f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.989438] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c33657f-29f9-4c38-80c7-0df3e08ea275 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.016141] env[61964]: DEBUG nova.compute.provider_tree [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.034761] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 780.035550] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-488c6627-7b53-4fa1-b815-27c9a84f363b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.050616] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 780.050616] env[61964]: value = "task-1040886" [ 780.050616] env[61964]: _type = "Task" [ 780.050616] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.062490] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040886, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.078196] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 780.078773] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 780.078773] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Deleting the datastore file [datastore2] af413b1d-9e7e-43d0-a39a-fb9dda9cf281 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.078953] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f7c9392-27f1-480f-ada2-14abd77cb4d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.087425] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for the task: (returnval){ [ 780.087425] env[61964]: value = "task-1040887" [ 780.087425] env[61964]: _type = "Task" [ 780.087425] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.100260] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.172447] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040884, 'name': ReconfigVM_Task, 'duration_secs': 0.151379} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.172782] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230373', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'name': 'volume-8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6ee845c3-5f79-4704-8b7b-cd3770202647', 'attached_at': '', 'detached_at': '', 'volume_id': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9', 'serial': '8352b6f4-47b3-40e9-93a1-3ce98bf146a9'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 780.173098] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 780.174028] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7418403-b90d-42d6-8054-f31249fbaf65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.183062] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 780.183369] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05d26525-455e-403d-89fe-5351084c9d5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.207226] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 780.243433] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 780.243433] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.243433] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.243646] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.243675] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.243825] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 780.244050] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 780.244289] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 780.244547] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 780.244722] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 780.245025] env[61964]: DEBUG nova.virt.hardware [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 780.245808] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166b5856-4d82-40d2-ae44-42bcb7d29ff8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.255848] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b2162c-9959-4711-9ac3-4fe7b248d9a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.276146] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 780.276597] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 780.276863] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Deleting the datastore file [datastore1] 6ee845c3-5f79-4704-8b7b-cd3770202647 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.277197] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-940b4d02-3b73-435e-9182-a5d795fe3de2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.286958] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for the task: (returnval){ [ 780.286958] env[61964]: value = "task-1040889" [ 780.286958] env[61964]: _type = "Task" [ 780.286958] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.295824] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.358338] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52136db8-d481-948d-a123-c5b02583d97e, 'name': SearchDatastore_Task, 'duration_secs': 0.017926} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.358939] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.359538] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 595c94f1-9dec-454d-b301-5656287f53de/595c94f1-9dec-454d-b301-5656287f53de.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.359945] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31e24e0b-d118-41d1-88e3-cdf788c11ac5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.373430] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 780.373430] env[61964]: value = "task-1040890" [ 780.373430] env[61964]: _type = "Task" [ 780.373430] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.388029] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040890, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.437518] env[61964]: DEBUG nova.compute.manager [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-changed-0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 780.437957] env[61964]: DEBUG nova.compute.manager [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing instance network info cache due to event network-changed-0db516e7-e7df-4355-b5d0-0d2dad89e921. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 780.438351] env[61964]: DEBUG oslo_concurrency.lockutils [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.438615] env[61964]: DEBUG oslo_concurrency.lockutils [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.438900] env[61964]: DEBUG nova.network.neutron [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing network info cache for port 0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.519933] env[61964]: DEBUG nova.scheduler.client.report [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.563769] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040886, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.602668] env[61964]: DEBUG oslo_vmware.api [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Task: {'id': task-1040887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289191} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.604615] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.604615] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.604615] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.604615] env[61964]: INFO nova.compute.manager [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Took 1.20 seconds to destroy the instance on the hypervisor. [ 780.604615] env[61964]: DEBUG oslo.service.loopingcall [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.605292] env[61964]: DEBUG nova.compute.manager [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 780.605347] env[61964]: DEBUG nova.network.neutron [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.804223] env[61964]: DEBUG oslo_vmware.api [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Task: {'id': task-1040889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152248} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.804223] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.804223] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.804223] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.804223] env[61964]: INFO nova.compute.manager [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Took 2.34 seconds to destroy the instance on the hypervisor. [ 780.804223] env[61964]: DEBUG oslo.service.loopingcall [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.804223] env[61964]: DEBUG nova.compute.manager [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 780.804223] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.901842] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040890, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.027708] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.855s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.028700] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 781.033928] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.390s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.036532] env[61964]: INFO nova.compute.claims [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.070396] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040886, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.387739] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040890, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.736047} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.389254] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 595c94f1-9dec-454d-b301-5656287f53de/595c94f1-9dec-454d-b301-5656287f53de.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 781.389254] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.389254] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e378c8ce-6eb8-4d2c-9500-a34bfdaf20be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.397152] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 781.397152] env[61964]: value = "task-1040891" [ 781.397152] env[61964]: _type = "Task" [ 781.397152] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.406426] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.545030] env[61964]: DEBUG nova.compute.utils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.551990] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 781.551990] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 781.568349] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040886, 'name': CloneVM_Task, 'duration_secs': 1.399796} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.568526] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created linked-clone VM from snapshot [ 781.571852] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368282a6-5fa5-4fec-a57b-6d31b98be9d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.582295] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploading image a4bc1cdf-2405-42f5-a7fb-757fd7700645 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 781.615621] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 781.615621] env[61964]: value = "vm-230465" [ 781.615621] env[61964]: _type = "VirtualMachine" [ 781.615621] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 781.615934] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-129365d8-72e0-49ec-9b50-203c4d4af1e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.626615] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease: (returnval){ [ 781.626615] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52023f41-bfe6-9ef5-b78e-586060cb828c" [ 781.626615] env[61964]: _type = "HttpNfcLease" [ 781.626615] env[61964]: } obtained for exporting VM: (result){ [ 781.626615] env[61964]: value = "vm-230465" [ 781.626615] env[61964]: _type = "VirtualMachine" [ 781.626615] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 781.626615] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the lease: (returnval){ [ 781.626615] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52023f41-bfe6-9ef5-b78e-586060cb828c" [ 781.626615] env[61964]: _type = "HttpNfcLease" [ 781.626615] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 781.635632] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 781.635632] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52023f41-bfe6-9ef5-b78e-586060cb828c" [ 781.635632] env[61964]: _type = "HttpNfcLease" [ 781.635632] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 781.639321] env[61964]: DEBUG nova.network.neutron [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updated VIF entry in instance network info cache for port 0db516e7-e7df-4355-b5d0-0d2dad89e921. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.639723] env[61964]: DEBUG nova.network.neutron [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.690382] env[61964]: DEBUG nova.policy [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f964ed45226a4c3690321775e8d3c100', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '787322c793384f849d0a2acae27a2052', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.747882] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Successfully updated port: 0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.918183] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.250977} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.918183] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.918183] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce0077c-96bc-4ada-b29b-38cf43bf410e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.948595] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 595c94f1-9dec-454d-b301-5656287f53de/595c94f1-9dec-454d-b301-5656287f53de.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.948595] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f2fa370-24a4-4794-b3f8-6b05f8a92128 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.972627] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 781.972627] env[61964]: value = "task-1040893" [ 781.972627] env[61964]: _type = "Task" [ 781.972627] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.983024] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.009274] env[61964]: DEBUG nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Received event network-vif-deleted-d1e722bd-f0de-4c54-b282-7ea69ab92fd0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 782.009742] env[61964]: INFO nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Neutron deleted interface d1e722bd-f0de-4c54-b282-7ea69ab92fd0; detaching it from the instance and deleting it from the info cache [ 782.010109] env[61964]: DEBUG nova.network.neutron [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.050148] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 782.140637] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 782.140637] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52023f41-bfe6-9ef5-b78e-586060cb828c" [ 782.140637] env[61964]: _type = "HttpNfcLease" [ 782.140637] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 782.141017] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 782.141017] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52023f41-bfe6-9ef5-b78e-586060cb828c" [ 782.141017] env[61964]: _type = "HttpNfcLease" [ 782.141017] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 782.141981] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be965a51-2140-40bf-bd06-e57778e365ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.145283] env[61964]: DEBUG oslo_concurrency.lockutils [req-e3c1b171-99a4-40e3-9515-84cb5b166e54 req-8c786163-7df9-43c7-80f8-56aed654d12f service nova] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.153686] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 782.153686] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 782.250687] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.250835] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.250991] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.263838] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-983bf841-3e2d-4a13-9d79-ff7d4b7e382e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.305114] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Successfully created port: 6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.469976] env[61964]: DEBUG nova.network.neutron [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.473828] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.484477] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040893, 'name': ReconfigVM_Task, 'duration_secs': 0.359621} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.487385] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 595c94f1-9dec-454d-b301-5656287f53de/595c94f1-9dec-454d-b301-5656287f53de.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.489029] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4c2c20a-1bfe-4cd3-981c-d497b6c972f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.501411] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 782.501411] env[61964]: value = "task-1040894" [ 782.501411] env[61964]: _type = "Task" [ 782.501411] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.512561] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3b0abf90-b640-4950-9c62-e0e1fbe7f94f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.518906] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040894, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.533707] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4cd257-33e9-4029-8875-935a8e2572e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.557912] env[61964]: DEBUG nova.compute.manager [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Received event network-vif-plugged-0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 782.557912] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Acquiring lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.557912] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.557912] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.557912] env[61964]: DEBUG nova.compute.manager [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] No waiting events found dispatching network-vif-plugged-0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 782.557912] env[61964]: WARNING nova.compute.manager [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Received unexpected event network-vif-plugged-0a8247ec-f054-4079-ad48-3b272687dce4 for instance with vm_state building and task_state spawning. [ 782.557912] env[61964]: DEBUG nova.compute.manager [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Received event network-changed-0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 782.558293] env[61964]: DEBUG nova.compute.manager [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Refreshing instance network info cache due to event network-changed-0a8247ec-f054-4079-ad48-3b272687dce4. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 782.558489] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Acquiring lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.581726] env[61964]: DEBUG nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Detach interface failed, port_id=d1e722bd-f0de-4c54-b282-7ea69ab92fd0, reason: Instance af413b1d-9e7e-43d0-a39a-fb9dda9cf281 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 782.581925] env[61964]: DEBUG nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Received event network-vif-deleted-ef8e5acc-8a42-44d0-84c4-57803733186b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 782.582145] env[61964]: INFO nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Neutron deleted interface ef8e5acc-8a42-44d0-84c4-57803733186b; detaching it from the instance and deleting it from the info cache [ 782.582321] env[61964]: DEBUG nova.network.neutron [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.806340] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.834032] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b83afa-59b1-466f-b172-0e7490e982e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.843847] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af41cd3-2d32-4e40-b9fa-fbe9c86c197e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.891703] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0131ab8-831c-4c7a-aa4f-73118088e46c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.904215] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21d4183-b874-4ac8-ab33-22159b17a994 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.926828] env[61964]: DEBUG nova.compute.provider_tree [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.974856] env[61964]: INFO nova.compute.manager [-] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Took 2.37 seconds to deallocate network for instance. [ 782.979262] env[61964]: INFO nova.compute.manager [-] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Took 2.18 seconds to deallocate network for instance. [ 783.024712] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040894, 'name': Rename_Task, 'duration_secs': 0.211351} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.028015] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 783.028477] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea707516-baef-4809-b0c8-e71ce6f4bc9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.043924] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 783.043924] env[61964]: value = "task-1040895" [ 783.043924] env[61964]: _type = "Task" [ 783.043924] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.058228] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.066644] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 783.090752] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d96d22e-841e-420f-80f0-08b2ba417a6d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.103846] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfd13cd-3d96-425d-a607-4aca1c5b5cc0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.120404] env[61964]: DEBUG nova.network.neutron [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Updating instance_info_cache with network_info: [{"id": "0a8247ec-f054-4079-ad48-3b272687dce4", "address": "fa:16:3e:03:7b:8c", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a8247ec-f0", "ovs_interfaceid": "0a8247ec-f054-4079-ad48-3b272687dce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.149335] env[61964]: DEBUG nova.compute.manager [req-50acba39-2c50-4f22-91e1-47d8d3ac4e1b req-0091c043-e80a-47e2-94b1-d0de00f93d7e service nova] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Detach interface failed, port_id=ef8e5acc-8a42-44d0-84c4-57803733186b, reason: Instance 6ee845c3-5f79-4704-8b7b-cd3770202647 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 783.431286] env[61964]: DEBUG nova.scheduler.client.report [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 783.494824] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.556113] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040895, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.583176] env[61964]: INFO nova.compute.manager [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Took 0.60 seconds to detach 1 volumes for instance. [ 783.586032] env[61964]: DEBUG nova.compute.manager [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Deleting volume: 8352b6f4-47b3-40e9-93a1-3ce98bf146a9 {{(pid=61964) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 783.625513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.625513] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Instance network_info: |[{"id": "0a8247ec-f054-4079-ad48-3b272687dce4", "address": "fa:16:3e:03:7b:8c", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a8247ec-f0", "ovs_interfaceid": "0a8247ec-f054-4079-ad48-3b272687dce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 783.625513] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Acquired lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.625513] env[61964]: DEBUG nova.network.neutron [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Refreshing network info cache for port 0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.626399] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:7b:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a8247ec-f054-4079-ad48-3b272687dce4', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.637011] env[61964]: DEBUG oslo.service.loopingcall [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.645889] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.647349] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3875a637-42b8-4620-96d5-44bb123d9571 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.677213] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.677213] env[61964]: value = "task-1040896" [ 783.677213] env[61964]: _type = "Task" [ 783.677213] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.687504] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040896, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.937274] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.903s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.937920] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 783.941895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.253s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.942713] env[61964]: DEBUG nova.objects.instance [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lazy-loading 'resources' on Instance uuid 85cd73df-a2f2-4f54-b984-832f9589a236 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.060592] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040895, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.158819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.189942] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040896, 'name': CreateVM_Task, 'duration_secs': 0.511902} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.190317] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.191247] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.191614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.192282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.192356] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-939d9315-f25a-4338-9e4b-352b82571bc1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.200272] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 784.200272] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db46f5-de63-286e-de07-e481e8973022" [ 784.200272] env[61964]: _type = "Task" [ 784.200272] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.213844] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db46f5-de63-286e-de07-e481e8973022, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.447540] env[61964]: DEBUG nova.compute.utils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.452421] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 784.452737] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 784.560025] env[61964]: DEBUG oslo_vmware.api [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040895, 'name': PowerOnVM_Task, 'duration_secs': 1.220423} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.563679] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 784.564926] env[61964]: INFO nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Took 9.06 seconds to spawn the instance on the hypervisor. [ 784.566916] env[61964]: DEBUG nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 784.568571] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cca6831-f195-4229-8b9d-94f1b196b409 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.589588] env[61964]: DEBUG nova.policy [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dbb0a78b0154ebcb7e5d54b5daecc8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a474c9bb101143999bb79396fe1c52de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.627941] env[61964]: DEBUG nova.compute.manager [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Received event network-vif-plugged-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 784.628197] env[61964]: DEBUG oslo_concurrency.lockutils [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] Acquiring lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.628916] env[61964]: DEBUG oslo_concurrency.lockutils [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.629123] env[61964]: DEBUG oslo_concurrency.lockutils [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.629304] env[61964]: DEBUG nova.compute.manager [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] No waiting events found dispatching network-vif-plugged-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 784.629476] env[61964]: WARNING nova.compute.manager [req-408cdd0f-ebc9-49e2-8741-b26b40dd9e85 req-31d09ac5-9404-4410-a130-864c92113101 service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Received unexpected event network-vif-plugged-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 for instance with vm_state building and task_state spawning. [ 784.712298] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db46f5-de63-286e-de07-e481e8973022, 'name': SearchDatastore_Task, 'duration_secs': 0.019514} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.715274] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.715510] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.715778] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.719024] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.719024] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.719201] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8866f8ae-124d-46a3-949c-06264fe4ab85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.731798] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.732625] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 784.733641] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc1bffc1-0369-4762-9cf5-b107d78538ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.745060] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 784.745060] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b1f6d5-9d8b-3bee-5186-6d28e819cdf7" [ 784.745060] env[61964]: _type = "Task" [ 784.745060] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.756819] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b1f6d5-9d8b-3bee-5186-6d28e819cdf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.859273] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Successfully updated port: 6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 784.878298] env[61964]: DEBUG nova.network.neutron [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Updated VIF entry in instance network info cache for port 0a8247ec-f054-4079-ad48-3b272687dce4. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.879096] env[61964]: DEBUG nova.network.neutron [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Updating instance_info_cache with network_info: [{"id": "0a8247ec-f054-4079-ad48-3b272687dce4", "address": "fa:16:3e:03:7b:8c", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a8247ec-f0", "ovs_interfaceid": "0a8247ec-f054-4079-ad48-3b272687dce4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.885021] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9894b1-65f4-409c-b158-0fef00230bca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.892227] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e615ab-d83f-4608-a1f3-3ba0d69d4e3b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.938101] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d8a1a5-2eed-41e5-86b5-ec6588279811 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.952129] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e11cd6-d287-4a39-858c-4319e1a49a26 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.958164] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 784.975027] env[61964]: DEBUG nova.compute.provider_tree [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.092737] env[61964]: INFO nova.compute.manager [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Took 43.96 seconds to build instance. [ 785.156287] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Successfully created port: dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.272024] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b1f6d5-9d8b-3bee-5186-6d28e819cdf7, 'name': SearchDatastore_Task, 'duration_secs': 0.015415} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.272024] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e40ae1c0-e548-4c98-98e1-b97f9f14bd2e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.281656] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 785.281656] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52000df9-b46c-5c19-ff86-af1140c76736" [ 785.281656] env[61964]: _type = "Task" [ 785.281656] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.301892] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52000df9-b46c-5c19-ff86-af1140c76736, 'name': SearchDatastore_Task, 'duration_secs': 0.011863} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.301892] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.301892] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d/6a03c9cc-bf79-4d60-98fd-48b2e5bc547d.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 785.301892] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06de5738-872b-4798-8339-4d77453269ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.312594] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 785.312594] env[61964]: value = "task-1040898" [ 785.312594] env[61964]: _type = "Task" [ 785.312594] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.328953] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.363633] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.363816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.364038] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.387323] env[61964]: DEBUG oslo_concurrency.lockutils [req-18e0379c-8c3a-4534-bf39-61594ca390b2 req-c6a4e6da-40b1-4c1d-962b-553f390fa073 service nova] Releasing lock "refresh_cache-6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.481049] env[61964]: DEBUG nova.scheduler.client.report [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.597580] env[61964]: DEBUG oslo_concurrency.lockutils [None req-84c75bcd-5c91-4b89-9f39-4a54233f2f45 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.567s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.751595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "595c94f1-9dec-454d-b301-5656287f53de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.751887] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.753180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "595c94f1-9dec-454d-b301-5656287f53de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.753180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.753180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.762895] env[61964]: INFO nova.compute.manager [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Terminating instance [ 785.827288] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040898, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479555} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.829407] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d/6a03c9cc-bf79-4d60-98fd-48b2e5bc547d.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.829407] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.829407] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a485289-721f-4b66-b608-1728724cdc65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.841649] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 785.841649] env[61964]: value = "task-1040899" [ 785.841649] env[61964]: _type = "Task" [ 785.841649] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.853481] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.919268] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.974171] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 785.986624] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.986624] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.993020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.998286] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 30.663s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.028727] env[61964]: INFO nova.scheduler.client.report [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Deleted allocations for instance 85cd73df-a2f2-4f54-b984-832f9589a236 [ 786.097124] env[61964]: DEBUG nova.network.neutron [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updating instance_info_cache with network_info: [{"id": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "address": "fa:16:3e:51:4c:fb", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e60e0a5-9e", "ovs_interfaceid": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.270042] env[61964]: DEBUG nova.compute.manager [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 786.270042] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 786.270718] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bcdd1c-cb8e-4ac9-a516-b73458f6abb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.279990] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 786.280642] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-861e4be4-8017-4b58-b126-fbce8675ccef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.288577] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 786.288577] env[61964]: value = "task-1040900" [ 786.288577] env[61964]: _type = "Task" [ 786.288577] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.299247] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.353038] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082316} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.353379] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.354248] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6585c86a-8cd6-4efa-bae0-d44215f24795 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.378494] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d/6a03c9cc-bf79-4d60-98fd-48b2e5bc547d.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.379286] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3877ea63-df1c-4ed4-9a7f-94276c84b104 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.402535] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 786.402535] env[61964]: value = "task-1040901" [ 786.402535] env[61964]: _type = "Task" [ 786.402535] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.416507] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040901, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.499155] env[61964]: DEBUG nova.compute.utils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.537421] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c27629d-7465-414e-ab95-60ecc0c84ba7 tempest-ServerMetadataNegativeTestJSON-1841818311 tempest-ServerMetadataNegativeTestJSON-1841818311-project-member] Lock "85cd73df-a2f2-4f54-b984-832f9589a236" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.495s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.601172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.601514] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Instance network_info: |[{"id": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "address": "fa:16:3e:51:4c:fb", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e60e0a5-9e", "ovs_interfaceid": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 786.749220] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Successfully updated port: dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.801739] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040900, 'name': PowerOffVM_Task, 'duration_secs': 0.2338} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.802108] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 786.802302] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 786.802570] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f39d570-e9e1-446c-b165-5ecfe0c178c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.882942] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 786.883231] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 786.883628] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleting the datastore file [datastore2] 595c94f1-9dec-454d-b301-5656287f53de {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.883956] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5cb9bfaf-840d-4fa0-b043-11a80a7be3b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.892216] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 786.892216] env[61964]: value = "task-1040903" [ 786.892216] env[61964]: _type = "Task" [ 786.892216] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.901458] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.912951] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040901, 'name': ReconfigVM_Task, 'duration_secs': 0.498171} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.913164] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d/6a03c9cc-bf79-4d60-98fd-48b2e5bc547d.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.913774] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b7965b0-fd52-4fc0-b30c-f62633dcbcc5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.921816] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 786.921816] env[61964]: value = "task-1040904" [ 786.921816] env[61964]: _type = "Task" [ 786.921816] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.931617] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040904, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.002821] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.040232] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 63793ce6-d511-403e-8a4b-cad8c4157449 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.040412] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a995ff8c-75b9-4089-ad4c-9e6baa33fe8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.040525] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.040642] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance aef4c3a7-641a-4356-9187-ae4c082ccde9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.040757] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 16276be9-b305-4d1d-afde-bc98be42687a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.040904] env[61964]: WARNING nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance af413b1d-9e7e-43d0-a39a-fb9dda9cf281 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.041034] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance df0a9a69-bd44-4da4-ba3a-9ba241c010a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.041166] env[61964]: WARNING nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6ee845c3-5f79-4704-8b7b-cd3770202647 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.041279] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 46435c68-f85a-4360-b2b5-6296afc33c3e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.041394] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 4104dee9-34db-4396-9189-8dc677d4b677 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.041505] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 58f8d368-fa0d-4c8d-9009-e6e1c5940032 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.041627] env[61964]: WARNING nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance fb538d67-47c3-4639-a642-b3358f97df31 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.041749] env[61964]: WARNING nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 0156727e-60dc-4e24-99e6-96b3e735aa0e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.041862] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.041973] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 9adedaa2-8594-4ddd-8f85-a41174207ef8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.042097] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance e93f98b3-4763-4f02-abd3-c24a9ab8dc17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.042209] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 595c94f1-9dec-454d-b301-5656287f53de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.042332] env[61964]: WARNING nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 65566c20-6d69-471c-b098-3c30c01d9955 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.042444] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.042552] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance beddafaf-8a91-4bfa-af3c-fa69977677ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.042662] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 062da016-f4d6-424d-90e6-5903940e2daa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.252394] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.252537] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.252732] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.406166] env[61964]: DEBUG oslo_vmware.api [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1040903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234344} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.406166] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.406166] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 787.406166] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 787.406300] env[61964]: INFO nova.compute.manager [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Took 1.14 seconds to destroy the instance on the hypervisor. [ 787.406492] env[61964]: DEBUG oslo.service.loopingcall [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.406791] env[61964]: DEBUG nova.compute.manager [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 787.406883] env[61964]: DEBUG nova.network.neutron [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 787.432914] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040904, 'name': Rename_Task, 'duration_secs': 0.223495} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.433224] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.433483] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-013435fa-43c8-48d9-8a45-fd46434a4656 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.442561] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 787.442561] env[61964]: value = "task-1040905" [ 787.442561] env[61964]: _type = "Task" [ 787.442561] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.452190] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.546367] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 560bd013-eae1-4541-aaa2-0732d8f20e4a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.791588] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 787.874035] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 787.874360] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 787.874360] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 787.874416] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 787.874592] env[61964]: DEBUG nova.virt.hardware [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 787.876856] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c16c8f-b815-45e2-938e-a31d2d1fce44 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.892158] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 787.892453] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.892640] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.892911] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.893169] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.893354] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 787.893572] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 787.893739] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 787.893912] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 787.894862] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 787.895125] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 787.896449] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59f6dfd-9145-447b-a58a-d93bb25830d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.903258] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c7e77d-7e9c-4304-84bf-915eb89ba17e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.914216] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 787.925139] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d247c9-ea2e-42ee-b61b-c519a01ce502 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.933020] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf5ec34-2b74-4db5-9ac1-959004b7e0fb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.933630] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:4c:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '412cde91-d0f0-4193-b36b-d8b9d17384c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e60e0a5-9e9f-43f3-8823-cc1d7c0be746', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.942213] env[61964]: DEBUG oslo.service.loopingcall [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.945462] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 787.946544] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd03a4f7-854c-42d3-af78-b0a247f76308 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.976181] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 787.976389] env[61964]: ERROR oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk due to incomplete transfer. [ 787.977755] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-69d2456d-ce6b-420f-b858-34cedbfd9fad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.983356] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040905, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.988220] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.988220] env[61964]: value = "task-1040906" [ 787.988220] env[61964]: _type = "Task" [ 787.988220] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.989132] env[61964]: DEBUG oslo_vmware.rw_handles [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ee1c-07cf-0c44-a827-366ba50e796a/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 787.989425] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Uploaded image 95e7d434-615c-44b6-a948-97f1713d7c9f to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 787.992502] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 787.993165] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1f954006-c020-467e-a53c-8274a408034c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.998917] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040906, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.001417] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 788.001417] env[61964]: value = "task-1040907" [ 788.001417] env[61964]: _type = "Task" [ 788.001417] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.011883] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040907, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.020638] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Updating instance_info_cache with network_info: [{"id": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "address": "fa:16:3e:fa:9d:dc", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc3f581d-2c", "ovs_interfaceid": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.050094] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a23e7f40-1b3e-4c8f-b664-30f6314923a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.083886] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.084331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.084646] env[61964]: INFO nova.compute.manager [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Attaching volume c8b09e3d-bd36-431c-9495-358374f38753 to /dev/sdb [ 788.178918] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a5e7e4-af21-46d8-a15d-314e804536e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.186410] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157d6ab1-c0d9-487b-b609-4e17de280a40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.204625] env[61964]: DEBUG nova.virt.block_device [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating existing volume attachment record: 0b175c48-03c5-45a1-98c3-de9265542ce4 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 788.226438] env[61964]: DEBUG nova.compute.manager [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Received event network-changed-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 788.226438] env[61964]: DEBUG nova.compute.manager [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Refreshing instance network info cache due to event network-changed-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 788.226438] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Acquiring lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.226438] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Acquired lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.226438] env[61964]: DEBUG nova.network.neutron [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Refreshing network info cache for port 6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.235559] env[61964]: DEBUG nova.network.neutron [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.457681] env[61964]: DEBUG oslo_vmware.api [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040905, 'name': PowerOnVM_Task, 'duration_secs': 0.610602} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.457989] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.458235] env[61964]: INFO nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Took 8.25 seconds to spawn the instance on the hypervisor. [ 788.458400] env[61964]: DEBUG nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 788.459279] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cf1e45-4b60-48bf-94f9-c0deadad824b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.499144] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040906, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.512261] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040907, 'name': Destroy_Task, 'duration_secs': 0.434204} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.513285] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Destroyed the VM [ 788.514920] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 788.515823] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3bf5d11d-c9ad-45d9-b2cc-ff227b3f1bc0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.524941] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.525390] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Instance network_info: |[{"id": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "address": "fa:16:3e:fa:9d:dc", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc3f581d-2c", "ovs_interfaceid": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 788.525676] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 788.525676] env[61964]: value = "task-1040909" [ 788.525676] env[61964]: _type = "Task" [ 788.525676] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.526493] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:9d:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc3f581d-2c02-4c24-a56e-e9e2923931cb', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.534960] env[61964]: DEBUG oslo.service.loopingcall [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.535795] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 788.539517] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-876c939a-fbea-441d-92a7-5d2d449c3e7e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.556274] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance ab33fb75-4114-4a26-8692-befb407b7460 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.565431] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040909, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.569167] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.569167] env[61964]: value = "task-1040911" [ 788.569167] env[61964]: _type = "Task" [ 788.569167] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.575486] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040911, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.738577] env[61964]: INFO nova.compute.manager [-] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Took 1.33 seconds to deallocate network for instance. [ 788.859436] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "4104dee9-34db-4396-9189-8dc677d4b677" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.859758] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.859995] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "4104dee9-34db-4396-9189-8dc677d4b677-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.860210] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.860387] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.863476] env[61964]: INFO nova.compute.manager [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Terminating instance [ 788.984084] env[61964]: INFO nova.compute.manager [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Took 41.57 seconds to build instance. [ 789.002181] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040906, 'name': CreateVM_Task, 'duration_secs': 0.565315} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.002447] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 789.003304] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.003822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.004462] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.004865] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c95979d4-b4fd-4ca6-aefa-060a828c0413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.012089] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 789.012089] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528284fd-ebce-e235-62f0-81f46df62e90" [ 789.012089] env[61964]: _type = "Task" [ 789.012089] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.023589] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528284fd-ebce-e235-62f0-81f46df62e90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.045841] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040909, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.061444] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 1c922228-4ba9-43ed-8e5c-3ca15e48ceab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.079464] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040911, 'name': CreateVM_Task, 'duration_secs': 0.486714} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.079924] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 789.080543] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.136194] env[61964]: DEBUG nova.network.neutron [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updated VIF entry in instance network info cache for port 6e60e0a5-9e9f-43f3-8823-cc1d7c0be746. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.136965] env[61964]: DEBUG nova.network.neutron [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updating instance_info_cache with network_info: [{"id": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "address": "fa:16:3e:51:4c:fb", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e60e0a5-9e", "ovs_interfaceid": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.140077] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.143451] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.143451] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.143451] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.143451] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.146011] env[61964]: INFO nova.compute.manager [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Terminating instance [ 789.247077] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.377267] env[61964]: DEBUG nova.compute.manager [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 789.377267] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 789.377267] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba014117-019a-4f33-a7f3-558e4a9002b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.389517] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 789.389517] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69d6d2a1-8b7c-4fb9-be70-a01a5bf6fe5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.403311] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 789.403311] env[61964]: value = "task-1040913" [ 789.403311] env[61964]: _type = "Task" [ 789.403311] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.414794] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.494762] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13c9a10-9cc7-4afc-9c7c-68f1f65ef189 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.207s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.523707] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528284fd-ebce-e235-62f0-81f46df62e90, 'name': SearchDatastore_Task, 'duration_secs': 0.012941} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.524091] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.524335] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.524637] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.525781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.525781] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.525781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.525781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.526015] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e24d557-70b2-413f-b4a0-b16cf5efa74c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.530021] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96cdb7ea-740c-44f3-94ba-4c1a7909663c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.536123] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 789.536123] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4e438-3765-f71f-4c41-a12c795d2a72" [ 789.536123] env[61964]: _type = "Task" [ 789.536123] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.545388] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.545657] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 789.546861] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fade83c8-8e1a-4fe7-b0df-e6ad53e2a806 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.557504] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4e438-3765-f71f-4c41-a12c795d2a72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.558083] env[61964]: DEBUG oslo_vmware.api [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040909, 'name': RemoveSnapshot_Task, 'duration_secs': 0.653555} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.558891] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 789.559230] env[61964]: INFO nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Took 14.96 seconds to snapshot the instance on the hypervisor. [ 789.564170] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 789.564170] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f2f137-abb2-e92c-d31f-453ace13466b" [ 789.564170] env[61964]: _type = "Task" [ 789.564170] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.572640] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 2a185e4b-c636-418a-a75e-2cf158f550c0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.574397] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 16 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 789.574639] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3584MB phys_disk=200GB used_disk=15GB total_vcpus=48 used_vcpus=16 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 789.589124] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f2f137-abb2-e92c-d31f-453ace13466b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.643384] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Releasing lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.643684] env[61964]: DEBUG nova.compute.manager [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Received event network-vif-plugged-dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 789.644228] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Acquiring lock "062da016-f4d6-424d-90e6-5903940e2daa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.644228] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Lock "062da016-f4d6-424d-90e6-5903940e2daa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.644331] env[61964]: DEBUG oslo_concurrency.lockutils [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] Lock "062da016-f4d6-424d-90e6-5903940e2daa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.644417] env[61964]: DEBUG nova.compute.manager [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] No waiting events found dispatching network-vif-plugged-dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 789.644592] env[61964]: WARNING nova.compute.manager [req-6cf9f47d-3d53-4934-9f28-6213592302fa req-13a7f9f9-71a1-4dc3-acce-cc14f6fab05a service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Received unexpected event network-vif-plugged-dc3f581d-2c02-4c24-a56e-e9e2923931cb for instance with vm_state building and task_state spawning. [ 789.654043] env[61964]: DEBUG nova.compute.manager [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 789.654299] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 789.655935] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64204d95-de15-402f-b6b4-8efe13073148 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.665061] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 789.668094] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8c7bc14-626c-4dbc-86cb-d3b3022780e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.738416] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 789.742280] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 789.742280] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore1] 9adedaa2-8594-4ddd-8f85-a41174207ef8 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.742280] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c07d44cb-3ceb-4ee3-a7fe-1a23813c5919 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.749268] env[61964]: DEBUG oslo_vmware.api [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 789.749268] env[61964]: value = "task-1040915" [ 789.749268] env[61964]: _type = "Task" [ 789.749268] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.763310] env[61964]: DEBUG oslo_vmware.api [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040915, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.927075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.927324] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.929161] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040913, 'name': PowerOffVM_Task, 'duration_secs': 0.23659} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.932168] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 789.932168] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 789.932168] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-feb40999-40be-4005-8d16-f977e7457ad7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.004889] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 790.005186] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 790.005385] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Deleting the datastore file [datastore2] 4104dee9-34db-4396-9189-8dc677d4b677 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.005675] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86981578-2411-47f4-9ad3-880474c843b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.018312] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for the task: (returnval){ [ 790.018312] env[61964]: value = "task-1040917" [ 790.018312] env[61964]: _type = "Task" [ 790.018312] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.031721] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.050456] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4e438-3765-f71f-4c41-a12c795d2a72, 'name': SearchDatastore_Task, 'duration_secs': 0.03993} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.050858] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.051017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.051283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.065861] env[61964]: DEBUG nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance disappeared during snapshot {{(pid=61964) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 790.078646] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f2f137-abb2-e92c-d31f-453ace13466b, 'name': SearchDatastore_Task, 'duration_secs': 0.045034} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.082610] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1213f063-482a-4e96-831a-cba0f579aa85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.087913] env[61964]: DEBUG nova.compute.manager [None req-107885b4-17b9-40ff-a82f-e5bf631cae1d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image not found during clean up 95e7d434-615c-44b6-a948-97f1713d7c9f {{(pid=61964) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 790.094021] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 790.094021] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525da629-53b5-1494-3589-f3045923b8ef" [ 790.094021] env[61964]: _type = "Task" [ 790.094021] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.102611] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525da629-53b5-1494-3589-f3045923b8ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.105818] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e72ac0-c65b-4be1-baa4-bbe5812b9fe6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.116027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c860d3b8-c109-42c8-b067-35f0365fad7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.151753] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb58814d-ae1a-42a2-b765-549c8c8d7aa4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.162238] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c4cb1d-5f19-48a5-ab75-23eb6fd81b36 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.182550] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.211563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.211842] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.212068] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.212281] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.212459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.214622] env[61964]: INFO nova.compute.manager [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Terminating instance [ 790.258671] env[61964]: DEBUG nova.compute.manager [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Received event network-changed-dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 790.258874] env[61964]: DEBUG nova.compute.manager [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Refreshing instance network info cache due to event network-changed-dc3f581d-2c02-4c24-a56e-e9e2923931cb. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 790.259105] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] Acquiring lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.259276] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] Acquired lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.259523] env[61964]: DEBUG nova.network.neutron [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Refreshing network info cache for port dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.266707] env[61964]: DEBUG oslo_vmware.api [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1040915, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168773} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.267476] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.267667] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.267850] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.268057] env[61964]: INFO nova.compute.manager [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Took 0.61 seconds to destroy the instance on the hypervisor. [ 790.268319] env[61964]: DEBUG oslo.service.loopingcall [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.268512] env[61964]: DEBUG nova.compute.manager [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 790.268606] env[61964]: DEBUG nova.network.neutron [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.430829] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 790.529412] env[61964]: DEBUG oslo_vmware.api [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Task: {'id': task-1040917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16275} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.529639] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.530044] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.530044] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.530258] env[61964]: INFO nova.compute.manager [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Took 1.16 seconds to destroy the instance on the hypervisor. [ 790.530504] env[61964]: DEBUG oslo.service.loopingcall [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.530681] env[61964]: DEBUG nova.compute.manager [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 790.530815] env[61964]: DEBUG nova.network.neutron [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.606285] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525da629-53b5-1494-3589-f3045923b8ef, 'name': SearchDatastore_Task, 'duration_secs': 0.011538} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.606365] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.606593] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/beddafaf-8a91-4bfa-af3c-fa69977677ba.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 790.608176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.608176] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.608176] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0adfd00-a256-45ae-a3e1-8716124e5d95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.611100] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b81cd01e-7a98-418e-b0a1-b1ed6a56dd3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.618579] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 790.618579] env[61964]: value = "task-1040918" [ 790.618579] env[61964]: _type = "Task" [ 790.618579] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.623655] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.623920] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.627630] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2808c5fc-02d0-42cf-8811-dc3314e3948a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.638836] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.639577] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 790.639577] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52600fb9-8621-375b-b872-de1852f5f67a" [ 790.639577] env[61964]: _type = "Task" [ 790.639577] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.650024] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52600fb9-8621-375b-b872-de1852f5f67a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.686885] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 790.720199] env[61964]: DEBUG nova.compute.manager [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 790.720199] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 790.720435] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a06a96d-8d8d-4793-8516-4f9addc52d66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.733360] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 790.734132] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-772f5e88-9731-4ce5-8d11-dba64b2bc680 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.744967] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 790.744967] env[61964]: value = "task-1040920" [ 790.744967] env[61964]: _type = "Task" [ 790.744967] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.761204] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.961452] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.069674] env[61964]: DEBUG nova.network.neutron [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.135224] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040918, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.151981] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52600fb9-8621-375b-b872-de1852f5f67a, 'name': SearchDatastore_Task, 'duration_secs': 0.01481} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.152920] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7855d2af-1f1f-4684-92c7-53572d2541bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.162188] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 791.162188] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5952e-daac-3231-d5ac-52bc5cc56050" [ 791.162188] env[61964]: _type = "Task" [ 791.162188] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.173910] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5952e-daac-3231-d5ac-52bc5cc56050, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.202409] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 791.202409] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.204s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.202409] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.759s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.203886] env[61964]: INFO nova.compute.claims [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.209399] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 791.209828] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Cleaning up deleted instances {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11670}} [ 791.261875] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040920, 'name': PowerOffVM_Task, 'duration_secs': 0.266037} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.262197] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 791.262375] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 791.262636] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d349809c-acd7-4e08-9ea7-980a9b4e8946 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.329796] env[61964]: DEBUG nova.network.neutron [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Updated VIF entry in instance network info cache for port dc3f581d-2c02-4c24-a56e-e9e2923931cb. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.333030] env[61964]: DEBUG nova.network.neutron [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Updating instance_info_cache with network_info: [{"id": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "address": "fa:16:3e:fa:9d:dc", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc3f581d-2c", "ovs_interfaceid": "dc3f581d-2c02-4c24-a56e-e9e2923931cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.342145] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 791.342500] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 791.342743] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore1] 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 791.343700] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f37572d9-68c9-46e1-b226-a26376c1c363 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.352059] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 791.352059] env[61964]: value = "task-1040922" [ 791.352059] env[61964]: _type = "Task" [ 791.352059] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.362408] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.572592] env[61964]: INFO nova.compute.manager [-] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Took 1.30 seconds to deallocate network for instance. [ 791.630173] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557717} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.630482] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/beddafaf-8a91-4bfa-af3c-fa69977677ba.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 791.630701] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.630963] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75b61ffa-29b1-415e-b8e1-cb954d8f1534 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.643022] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 791.643022] env[61964]: value = "task-1040923" [ 791.643022] env[61964]: _type = "Task" [ 791.643022] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.654025] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.674397] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5952e-daac-3231-d5ac-52bc5cc56050, 'name': SearchDatastore_Task, 'duration_secs': 0.018934} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.674709] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.674984] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 062da016-f4d6-424d-90e6-5903940e2daa/062da016-f4d6-424d-90e6-5903940e2daa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.675462] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c4f5b08-13ef-4f3a-8cd2-b891917e259e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.683825] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 791.683825] env[61964]: value = "task-1040924" [ 791.683825] env[61964]: _type = "Task" [ 791.683825] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.694060] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.723675] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] There are 14 instances to clean {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11679}} [ 791.723964] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 85cd73df-a2f2-4f54-b984-832f9589a236] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 791.833723] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] Releasing lock "refresh_cache-062da016-f4d6-424d-90e6-5903940e2daa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.834132] env[61964]: DEBUG nova.compute.manager [req-aeccc87c-68d9-4701-bff8-7c7dc681cf32 req-25559ca3-54a6-47c4-8005-65710366a60c service nova] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Received event network-vif-deleted-63bb7733-05d2-4984-bbe6-57940384a3a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 791.862927] env[61964]: DEBUG oslo_vmware.api [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1040922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192767} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.863223] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 791.863414] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 791.863599] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 791.863777] env[61964]: INFO nova.compute.manager [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 791.864047] env[61964]: DEBUG oslo.service.loopingcall [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.864416] env[61964]: DEBUG nova.compute.manager [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 791.864507] env[61964]: DEBUG nova.network.neutron [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 792.086629] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.151629] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125696} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.152766] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.153063] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8ac02b-5c57-4071-ae4c-376bdbc23e09 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.179944] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/beddafaf-8a91-4bfa-af3c-fa69977677ba.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.184019] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60a435d6-296a-4c74-9c5b-5b09bf23206c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.207944] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51847} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.209679] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 062da016-f4d6-424d-90e6-5903940e2daa/062da016-f4d6-424d-90e6-5903940e2daa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.209933] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.210642] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 792.210642] env[61964]: value = "task-1040925" [ 792.210642] env[61964]: _type = "Task" [ 792.210642] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.210642] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20b9eb3c-454c-42ab-bbeb-0836927c6ba5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.222264] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040925, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.226195] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 792.226195] env[61964]: value = "task-1040926" [ 792.226195] env[61964]: _type = "Task" [ 792.226195] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.228856] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: b50e683f-79b2-419a-ac95-94908adb2b3e] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 792.238305] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040926, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.326249] env[61964]: DEBUG nova.network.neutron [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.391534] env[61964]: DEBUG nova.compute.manager [req-c1f77fea-d14d-4153-adae-d73930f1adfd req-d8ebcda9-ea86-4c80-b4a6-48d7ab8d50c8 service nova] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Received event network-vif-deleted-4e7e78dc-a5cc-4704-aefd-8382de0f2c4f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 792.391746] env[61964]: DEBUG nova.compute.manager [req-c1f77fea-d14d-4153-adae-d73930f1adfd req-d8ebcda9-ea86-4c80-b4a6-48d7ab8d50c8 service nova] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Received event network-vif-deleted-ee1797a0-610e-49f9-9adc-a6ddcb0857a8 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 792.673363] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832b4b62-cede-4487-87f3-7ab34d00ebee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.683338] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb2e9f8-b57b-4fd1-8bc4-4ed5de0e8832 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.723947] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 792.728283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2238db-14f0-4098-8f76-60521773a70f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.731931] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b407c77-15ff-4f93-87cd-6e3eae6dfa90 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.740466] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e5e7b687-5691-4b1e-b113-2b8e83b23662] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 792.752921] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040925, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.755985] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc744c4-dae0-4fe5-a3af-ba77cdf1154d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.760782] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 792.761190] env[61964]: ERROR oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk due to incomplete transfer. [ 792.764458] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c9252cd3-b87e-4e28-bc08-f87160d0d46b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.766640] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040926, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089325} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.767841] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 792.768145] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230470', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'name': 'volume-c8b09e3d-bd36-431c-9495-358374f38753', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a995ff8c-75b9-4089-ad4c-9e6baa33fe8c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'serial': 'c8b09e3d-bd36-431c-9495-358374f38753'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 792.768971] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.769926] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e315df8-97a7-417c-9d63-ee63b323fd0f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.782380] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872a6ac5-ba44-438c-8aa7-3a1c11be3070 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.785652] env[61964]: DEBUG nova.compute.provider_tree [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.789410] env[61964]: DEBUG oslo_vmware.rw_handles [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cc492b-4146-529d-17e2-186dfcd70f0e/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 792.789668] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploaded image a4bc1cdf-2405-42f5-a7fb-757fd7700645 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 792.791431] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 792.812563] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fcdbd5bb-55bc-461f-af85-4274f764dd0c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.825043] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 062da016-f4d6-424d-90e6-5903940e2daa/062da016-f4d6-424d-90e6-5903940e2daa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.826157] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b512ee4e-e033-4067-8580-8587490ef9b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.829235] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d79f11a1-e452-4c21-a04c-78deac3c8def {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.844529] env[61964]: INFO nova.compute.manager [-] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Took 2.31 seconds to deallocate network for instance. [ 792.867881] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 792.867881] env[61964]: value = "task-1040927" [ 792.867881] env[61964]: _type = "Task" [ 792.867881] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.876110] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] volume-c8b09e3d-bd36-431c-9495-358374f38753/volume-c8b09e3d-bd36-431c-9495-358374f38753.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.878244] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f164ae34-ed1a-49cf-b162-181484b8460c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.892632] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 792.892632] env[61964]: value = "task-1040928" [ 792.892632] env[61964]: _type = "Task" [ 792.892632] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.903699] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 792.903699] env[61964]: value = "task-1040929" [ 792.903699] env[61964]: _type = "Task" [ 792.903699] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.907554] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040927, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.913845] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040928, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.920421] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040929, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.959685] env[61964]: DEBUG nova.network.neutron [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.238440] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040925, 'name': ReconfigVM_Task, 'duration_secs': 0.548885} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.238906] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfigured VM instance instance-0000003c to attach disk [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/beddafaf-8a91-4bfa-af3c-fa69977677ba.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.240095] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9cd03440-9367-4eef-9cd2-a128f5f18abc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.250582] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 793.250582] env[61964]: value = "task-1040930" [ 793.250582] env[61964]: _type = "Task" [ 793.250582] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.256007] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0ea19f52-b700-405f-8c3c-fbaa3e17b36a] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 793.267740] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040930, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.290632] env[61964]: DEBUG nova.scheduler.client.report [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.378902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.390398] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040927, 'name': Destroy_Task, 'duration_secs': 0.480979} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.390704] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroyed the VM [ 793.390947] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 793.391295] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-09c16b7a-f11f-4761-8ab0-aa56e172fc4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.408253] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 793.408253] env[61964]: value = "task-1040931" [ 793.408253] env[61964]: _type = "Task" [ 793.408253] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.415875] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040928, 'name': ReconfigVM_Task, 'duration_secs': 0.436368} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.419739] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 062da016-f4d6-424d-90e6-5903940e2daa/062da016-f4d6-424d-90e6-5903940e2daa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.420888] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b2faf2b-c1ba-43e0-b71b-b8f2318ead10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.427312] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040929, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.431751] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 793.431751] env[61964]: value = "task-1040932" [ 793.431751] env[61964]: _type = "Task" [ 793.431751] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.432012] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040931, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.449483] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040932, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.462233] env[61964]: INFO nova.compute.manager [-] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Took 1.60 seconds to deallocate network for instance. [ 793.762411] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: df62ed65-0a89-4f04-9b5a-f5f4214e6bd5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 793.764863] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040930, 'name': Rename_Task, 'duration_secs': 0.227137} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.765924] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.766860] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c34055f4-3fc2-4d5f-9fc6-b14b0ffbedf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.777210] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 793.777210] env[61964]: value = "task-1040933" [ 793.777210] env[61964]: _type = "Task" [ 793.777210] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.788314] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.801582] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.801582] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 793.802027] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.578s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.804526] env[61964]: INFO nova.compute.claims [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.930976] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040931, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.934818] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040929, 'name': ReconfigVM_Task, 'duration_secs': 0.542227} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.938686] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfigured VM instance instance-00000021 to attach disk [datastore1] volume-c8b09e3d-bd36-431c-9495-358374f38753/volume-c8b09e3d-bd36-431c-9495-358374f38753.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.944275] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0d5c2f0-e7e0-438b-9e0f-2081df82fe8e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.962211] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040932, 'name': Rename_Task, 'duration_secs': 0.183561} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.963819] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.964175] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 793.964175] env[61964]: value = "task-1040934" [ 793.964175] env[61964]: _type = "Task" [ 793.964175] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.964535] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40e9c416-5781-44c2-be30-f96341d737e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.971816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.978558] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040934, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.980354] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 793.980354] env[61964]: value = "task-1040935" [ 793.980354] env[61964]: _type = "Task" [ 793.980354] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.993557] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.267865] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 5c340420-969b-4ccb-9f9a-7a833d8f0c43] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 794.293683] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040933, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.315293] env[61964]: DEBUG nova.compute.utils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.318360] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 794.318614] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.390150] env[61964]: DEBUG nova.policy [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dbb0a78b0154ebcb7e5d54b5daecc8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a474c9bb101143999bb79396fe1c52de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.423419] env[61964]: DEBUG nova.compute.manager [req-b4db99d7-e86c-402d-9b4c-1c523cefe509 req-2b1f5f92-209b-4263-83ae-5e85be52ce19 service nova] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Received event network-vif-deleted-0a8247ec-f054-4079-ad48-3b272687dce4 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 794.430961] env[61964]: DEBUG oslo_vmware.api [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040931, 'name': RemoveSnapshot_Task, 'duration_secs': 0.610478} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.431444] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 794.431659] env[61964]: INFO nova.compute.manager [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 16.51 seconds to snapshot the instance on the hypervisor. [ 794.477714] env[61964]: DEBUG oslo_vmware.api [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040934, 'name': ReconfigVM_Task, 'duration_secs': 0.1957} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.478107] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230470', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'name': 'volume-c8b09e3d-bd36-431c-9495-358374f38753', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a995ff8c-75b9-4089-ad4c-9e6baa33fe8c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'serial': 'c8b09e3d-bd36-431c-9495-358374f38753'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 794.499094] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040935, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.771744] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a2a25787-1426-4c09-892d-395e1becccc9] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 794.778478] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Successfully created port: 7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.794106] env[61964]: DEBUG oslo_vmware.api [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040933, 'name': PowerOnVM_Task, 'duration_secs': 0.5356} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.794295] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 794.794613] env[61964]: INFO nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Took 11.73 seconds to spawn the instance on the hypervisor. [ 794.794900] env[61964]: DEBUG nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 794.795996] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e5521e-165a-4a09-bb46-298710c73c41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.821646] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 794.995633] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040935, 'name': PowerOnVM_Task, 'duration_secs': 0.618455} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.001030] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 795.001030] env[61964]: INFO nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Took 9.03 seconds to spawn the instance on the hypervisor. [ 795.001030] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 795.001030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec686cbf-79cc-4e71-91a4-b374e8279b4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.019747] env[61964]: DEBUG nova.compute.manager [None req-132b7518-1b27-47f0-ae88-a9a730e31d17 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Found 1 images (rotation: 2) {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4985}} [ 795.274845] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6ced6800-db29-4766-8d83-b63b50d5fcc5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 795.308761] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfee9cf-95c6-4d08-a68d-6919658c94a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.327453] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaf1946-f6cb-44ff-ab0a-cf32ec094754 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.337346] env[61964]: INFO nova.compute.manager [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Took 46.37 seconds to build instance. [ 795.365581] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61151ba-9ea0-4bd7-9657-43f6a3a7bf08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.375266] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584878af-6520-42b2-9946-d87c56ef177d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.395035] env[61964]: DEBUG nova.compute.provider_tree [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.481613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.481613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.481613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.481613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.481613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.482743] env[61964]: INFO nova.compute.manager [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Terminating instance [ 795.532551] env[61964]: INFO nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Took 44.90 seconds to build instance. [ 795.545444] env[61964]: DEBUG nova.objects.instance [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lazy-loading 'flavor' on Instance uuid a995ff8c-75b9-4089-ad4c-9e6baa33fe8c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.779627] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: d99859dc-2206-42d9-ae6d-8294fbd6942c] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 795.839167] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 795.841935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc9c6628-ba17-4335-b3bb-34a929c6b9bc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.342s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.863539] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 795.863784] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.863951] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 795.864370] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.864582] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 795.864776] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 795.865058] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 795.865310] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 795.865538] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 795.865745] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 795.865962] env[61964]: DEBUG nova.virt.hardware [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 795.866913] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd9cb42-92fa-42e3-af1c-766df17436cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.876258] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db76ff3-6f47-4977-9c5a-d67a990ca486 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.897496] env[61964]: DEBUG nova.scheduler.client.report [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.965710] env[61964]: INFO nova.compute.manager [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Rescuing [ 795.966017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.966248] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.966408] env[61964]: DEBUG nova.network.neutron [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 795.986912] env[61964]: DEBUG nova.compute.manager [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 795.987729] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.988316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fdaad7-31ac-44e8-bbaa-b6c0d3c0ad4f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.996770] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 795.997050] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-435fa22d-1bc3-4cfe-9983-079647488e56 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.005554] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 796.005554] env[61964]: value = "task-1040936" [ 796.005554] env[61964]: _type = "Task" [ 796.005554] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.016085] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.034681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.363s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.051285] env[61964]: DEBUG oslo_concurrency.lockutils [None req-532ce302-8cfa-4f93-82ac-6a4292ade44f tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.967s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.282425] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 06631c59-9969-4763-bfe8-472e8dcf8848] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 796.403578] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.404154] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 796.407028] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.086s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.407301] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.410521] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.212s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.412059] env[61964]: INFO nova.compute.claims [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.427128] env[61964]: DEBUG nova.compute.manager [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 796.428802] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b89c120-52c2-4fe1-89d1-9ef326a05cf4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.445835] env[61964]: INFO nova.scheduler.client.report [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Deleted allocations for instance fb538d67-47c3-4639-a642-b3358f97df31 [ 796.518620] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040936, 'name': PowerOffVM_Task, 'duration_secs': 0.219495} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.518906] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 796.519092] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 796.519392] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3302cd77-7bea-4c00-8410-a9db9e6246ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.591767] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 796.592124] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 796.592331] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Deleting the datastore file [datastore1] 58f8d368-fa0d-4c8d-9009-e6e1c5940032 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.592617] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f707abb7-4b97-4b5b-8eaa-bf2b849f9d24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.600940] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for the task: (returnval){ [ 796.600940] env[61964]: value = "task-1040938" [ 796.600940] env[61964]: _type = "Task" [ 796.600940] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.611507] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040938, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.665802] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.666331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.787550] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0f1ed823-1531-4505-a68a-cb172f804655] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 796.892582] env[61964]: DEBUG nova.network.neutron [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updating instance_info_cache with network_info: [{"id": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "address": "fa:16:3e:51:4c:fb", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e60e0a5-9e", "ovs_interfaceid": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.898545] env[61964]: DEBUG nova.compute.manager [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Received event network-vif-plugged-7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 796.898711] env[61964]: DEBUG oslo_concurrency.lockutils [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] Acquiring lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.899055] env[61964]: DEBUG oslo_concurrency.lockutils [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.899246] env[61964]: DEBUG oslo_concurrency.lockutils [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.903026] env[61964]: DEBUG nova.compute.manager [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] No waiting events found dispatching network-vif-plugged-7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.903026] env[61964]: WARNING nova.compute.manager [req-f3803d9f-aa14-4f9d-bc9f-0a3eb1c21beb req-5f241d88-c76a-404f-92f5-af62d8f89ea3 service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Received unexpected event network-vif-plugged-7ac45fc2-5aeb-4136-85ad-abb23c24b047 for instance with vm_state building and task_state spawning. [ 796.916513] env[61964]: DEBUG nova.compute.utils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.921163] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 796.921435] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.942296] env[61964]: INFO nova.compute.manager [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] instance snapshotting [ 796.942890] env[61964]: DEBUG nova.objects.instance [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.963816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2dac315a-ae65-4071-823e-a8f306caa9d9 tempest-ServerPasswordTestJSON-1474672 tempest-ServerPasswordTestJSON-1474672-project-member] Lock "fb538d67-47c3-4639-a642-b3358f97df31" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.214s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.984422] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Successfully updated port: 7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.991471] env[61964]: DEBUG nova.policy [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f438e739578a4fabad898ebd5e2925cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ccf1fe5003e484fa73ce5c32767479f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.114283] env[61964]: DEBUG oslo_vmware.api [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Task: {'id': task-1040938, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165482} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.114656] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.115058] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.115452] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.115698] env[61964]: INFO nova.compute.manager [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Took 1.13 seconds to destroy the instance on the hypervisor. [ 797.116166] env[61964]: DEBUG oslo.service.loopingcall [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.116492] env[61964]: DEBUG nova.compute.manager [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.116847] env[61964]: DEBUG nova.network.neutron [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.169395] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 797.289850] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.290195] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.291787] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 9f216f01-dc9f-459f-88be-3a5439d3701a] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 797.398016] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.425930] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 797.450956] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d401ef9e-b5c2-45fc-a87e-4eb5ea49827a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.478652] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f8d052-7b8d-4943-ae37-e0fe53dcd487 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.488478] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.488629] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.488793] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.694481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.752018] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Successfully created port: 24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.797386] env[61964]: INFO nova.compute.manager [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Detaching volume c8b09e3d-bd36-431c-9495-358374f38753 [ 797.799727] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: d8d12f21-5af8-4156-8e89-3f719199ca0a] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 797.849463] env[61964]: INFO nova.virt.block_device [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Attempting to driver detach volume c8b09e3d-bd36-431c-9495-358374f38753 from mountpoint /dev/sdb [ 797.849463] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 797.849463] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230470', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'name': 'volume-c8b09e3d-bd36-431c-9495-358374f38753', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a995ff8c-75b9-4089-ad4c-9e6baa33fe8c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'serial': 'c8b09e3d-bd36-431c-9495-358374f38753'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 797.849463] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867176a8-9cfd-402a-b3be-cd9107922cb6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.880991] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad53e0c-83b6-49a5-ad4f-bc7e3e1b461c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.892101] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8474555-ee50-4126-ae2f-18390d0175c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.926622] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25b72ac-80ee-47f6-801d-e5a3a4290802 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.927598] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6526ae-202e-4cc3-962b-505066cc5518 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.952895] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] The volume has not been displaced from its original location: [datastore1] volume-c8b09e3d-bd36-431c-9495-358374f38753/volume-c8b09e3d-bd36-431c-9495-358374f38753.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 797.959816] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfiguring VM instance instance-00000021 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 797.962544] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7080ba7-09d8-4793-9157-cc2473421c84 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.976251] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81665462-0c7f-4991-8373-e8a16e8c43d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.011809] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 798.015346] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c93063cb-e1c3-4bdd-a03d-7d657b576f6f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.017820] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699a9c90-4b4d-4867-b36e-f10312e475e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.020814] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 798.020814] env[61964]: value = "task-1040939" [ 798.020814] env[61964]: _type = "Task" [ 798.020814] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.029378] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d623bd-1d3f-42a7-8792-a73d02e9459f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.034595] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 798.034595] env[61964]: value = "task-1040940" [ 798.034595] env[61964]: _type = "Task" [ 798.034595] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.038537] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040939, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.052309] env[61964]: DEBUG nova.compute.provider_tree [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.059053] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040940, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.081986] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.304043] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 4e14a519-b543-4a18-9e1d-7df16abf5345] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 798.447663] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 798.470324] env[61964]: DEBUG nova.network.neutron [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Updating instance_info_cache with network_info: [{"id": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "address": "fa:16:3e:e6:c2:eb", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ac45fc2-5a", "ovs_interfaceid": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.482811] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 798.484117] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 798.484439] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 798.484439] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 798.485339] env[61964]: DEBUG nova.virt.hardware [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 798.485765] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f026e2d-2398-4b29-adc6-e353641565c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.495717] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdaf6de-5e14-43a2-915f-70e1e8025848 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.531366] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040939, 'name': ReconfigVM_Task, 'duration_secs': 0.331042} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.531725] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Reconfigured VM instance instance-00000021 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 798.538213] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b5c6fa5-7f9b-4457-b6f0-fd9f11c9ca3f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.556091] env[61964]: DEBUG nova.scheduler.client.report [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 798.563359] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 798.563359] env[61964]: value = "task-1040941" [ 798.563359] env[61964]: _type = "Task" [ 798.563359] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.563545] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040940, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.572833] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040941, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.810679] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 798.810857] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Cleaning up deleted instances with incomplete migration {{(pid=61964) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11708}} [ 798.862927] env[61964]: DEBUG nova.network.neutron [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.931287] env[61964]: DEBUG nova.compute.manager [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Received event network-changed-7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 798.931456] env[61964]: DEBUG nova.compute.manager [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Refreshing instance network info cache due to event network-changed-7ac45fc2-5aeb-4136-85ad-abb23c24b047. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 798.931650] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] Acquiring lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.958024] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 798.958024] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c215f38-1ea2-4496-b378-dbb4e7037dda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.964929] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 798.964929] env[61964]: value = "task-1040942" [ 798.964929] env[61964]: _type = "Task" [ 798.964929] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.975293] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.975702] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Instance network_info: |[{"id": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "address": "fa:16:3e:e6:c2:eb", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ac45fc2-5a", "ovs_interfaceid": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 798.976196] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040942, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.976849] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] Acquired lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.977153] env[61964]: DEBUG nova.network.neutron [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Refreshing network info cache for port 7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.978801] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:c2:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ac45fc2-5aeb-4136-85ad-abb23c24b047', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.989485] env[61964]: DEBUG oslo.service.loopingcall [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.992869] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.993397] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70cb114a-ab5c-4cbc-a3fb-51500f41ced8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.017727] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.017727] env[61964]: value = "task-1040943" [ 799.017727] env[61964]: _type = "Task" [ 799.017727] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.030902] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040943, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.061462] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040940, 'name': CreateSnapshot_Task, 'duration_secs': 0.978797} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.061743] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 799.062565] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9fc417-74a0-4968-884e-efd7f102bd16 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.065662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.066159] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 799.070143] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.332s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.071585] env[61964]: INFO nova.compute.claims [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.087604] env[61964]: DEBUG oslo_vmware.api [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1040941, 'name': ReconfigVM_Task, 'duration_secs': 0.178093} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.088205] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230470', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'name': 'volume-c8b09e3d-bd36-431c-9495-358374f38753', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a995ff8c-75b9-4089-ad4c-9e6baa33fe8c', 'attached_at': '', 'detached_at': '', 'volume_id': 'c8b09e3d-bd36-431c-9495-358374f38753', 'serial': 'c8b09e3d-bd36-431c-9495-358374f38753'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 799.266558] env[61964]: DEBUG nova.network.neutron [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Updated VIF entry in instance network info cache for port 7ac45fc2-5aeb-4136-85ad-abb23c24b047. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.266935] env[61964]: DEBUG nova.network.neutron [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Updating instance_info_cache with network_info: [{"id": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "address": "fa:16:3e:e6:c2:eb", "network": {"id": "b4bbbb85-04e5-41db-8b55-e4fc5e79ca33", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-568574655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a474c9bb101143999bb79396fe1c52de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ac45fc2-5a", "ovs_interfaceid": "7ac45fc2-5aeb-4136-85ad-abb23c24b047", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.314127] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 799.369020] env[61964]: INFO nova.compute.manager [-] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Took 2.25 seconds to deallocate network for instance. [ 799.474767] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040942, 'name': PowerOffVM_Task, 'duration_secs': 0.416191} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.475056] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 799.475905] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24407b97-6187-481d-8195-8ba715d363f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.494425] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a50bc7b-3141-4c83-95ec-0ad7befa9659 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.528417] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040943, 'name': CreateVM_Task, 'duration_secs': 0.46002} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.530378] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 799.530582] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 799.530809] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c788b3a-381c-474e-a6ea-4abd2bd079f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.532756] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.532929] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.533260] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.533500] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc2566fc-744f-4935-a5ef-76f26667f2e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.538102] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 799.538102] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529f318d-6e92-b438-f294-25baf73b8049" [ 799.538102] env[61964]: _type = "Task" [ 799.538102] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.542138] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 799.542138] env[61964]: value = "task-1040944" [ 799.542138] env[61964]: _type = "Task" [ 799.542138] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.548678] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529f318d-6e92-b438-f294-25baf73b8049, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.553618] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 799.553724] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.553968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.554134] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.554316] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.554549] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a1bfcb2-010b-4f54-918f-5f3719b15a36 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.561733] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.561918] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.562630] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5122f9-2061-40c0-b555-b08d35179dbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.567636] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 799.567636] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521db33c-bbee-c698-6e07-a5d8ff44bc32" [ 799.567636] env[61964]: _type = "Task" [ 799.567636] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.574640] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521db33c-bbee-c698-6e07-a5d8ff44bc32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.580102] env[61964]: DEBUG nova.compute.utils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.583245] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 799.583416] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.597778] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 799.599494] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0e3eebb5-2380-4804-8801-6ec5b4f6bf9e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.610250] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 799.610250] env[61964]: value = "task-1040945" [ 799.610250] env[61964]: _type = "Task" [ 799.610250] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.617428] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040945, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.659019] env[61964]: DEBUG nova.objects.instance [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lazy-loading 'flavor' on Instance uuid a995ff8c-75b9-4089-ad4c-9e6baa33fe8c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.665301] env[61964]: DEBUG nova.policy [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '576d1881b73b43958ebbab05d8e523ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '540083b7cc4a4b8aa25714afc2add861', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.770707] env[61964]: DEBUG oslo_concurrency.lockutils [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] Releasing lock "refresh_cache-560bd013-eae1-4541-aaa2-0732d8f20e4a" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.770707] env[61964]: DEBUG nova.compute.manager [req-baa690d8-86ba-4970-8bb9-e8cadcb5b9b1 req-75597481-3ffa-4732-92e0-1bf08ab7337b service nova] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Received event network-vif-deleted-3038d60e-1ce8-479d-ac96-8b597d6d8d00 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 799.875066] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.968342] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Successfully updated port: 24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.050150] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529f318d-6e92-b438-f294-25baf73b8049, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.051273] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.051273] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.051273] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.081673] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521db33c-bbee-c698-6e07-a5d8ff44bc32, 'name': SearchDatastore_Task, 'duration_secs': 0.008239} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.082692] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaed5f60-aa8b-4982-a14c-8133ee7f662d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.087483] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 800.100902] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 800.100902] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52707fbb-cf33-0719-7664-aeaaa982e697" [ 800.100902] env[61964]: _type = "Task" [ 800.100902] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.110535] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52707fbb-cf33-0719-7664-aeaaa982e697, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.120567] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040945, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.139729] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Successfully created port: 1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.470467] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.470731] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.470772] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.472408] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f63a01-a75c-4634-99b0-470ed1b70a5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.480879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cf9944-0f72-4202-a6f5-559a81943dcf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.516035] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781ae244-7ac3-432b-b1f6-cbb87745532b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.524088] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd3063c-e97c-4b96-a38e-c18dd7a4851c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.537105] env[61964]: DEBUG nova.compute.provider_tree [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.611066] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52707fbb-cf33-0719-7664-aeaaa982e697, 'name': SearchDatastore_Task, 'duration_secs': 0.009718} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.614519] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.614828] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. {{(pid=61964) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 800.615625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.616027] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.616109] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fdce539-0947-4051-bdbe-4ccbcf9c9338 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.618931] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fffe4c12-02d3-41cd-a7bf-c1aded62ed19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.624183] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040945, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.628971] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 800.628971] env[61964]: value = "task-1040946" [ 800.628971] env[61964]: _type = "Task" [ 800.628971] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.633115] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.633115] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 800.634425] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78a16bf1-e1b3-4140-b30b-dc2af5525e5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.640492] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.645082] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 800.645082] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257d421-be9b-27fe-36b8-b4231305f678" [ 800.645082] env[61964]: _type = "Task" [ 800.645082] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.655601] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257d421-be9b-27fe-36b8-b4231305f678, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.665359] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38e7d7a0-ba20-4ee3-87a1-23626171f218 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.375s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.036167] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.040699] env[61964]: DEBUG nova.scheduler.client.report [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.054088] env[61964]: DEBUG nova.compute.manager [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Received event network-vif-plugged-24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 801.054088] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Acquiring lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.054088] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.054088] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.054088] env[61964]: DEBUG nova.compute.manager [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] No waiting events found dispatching network-vif-plugged-24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 801.054759] env[61964]: WARNING nova.compute.manager [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Received unexpected event network-vif-plugged-24e07767-885b-4cd5-9127-da8635ce0092 for instance with vm_state building and task_state spawning. [ 801.054759] env[61964]: DEBUG nova.compute.manager [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Received event network-changed-24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 801.056053] env[61964]: DEBUG nova.compute.manager [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Refreshing instance network info cache due to event network-changed-24e07767-885b-4cd5-9127-da8635ce0092. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 801.056053] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Acquiring lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.103264] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 801.138775] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040945, 'name': CloneVM_Task, 'duration_secs': 1.513444} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.141099] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 801.141338] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.141527] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.141719] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.141869] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.142026] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 801.142235] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 801.142397] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 801.142583] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 801.142763] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 801.142952] env[61964]: DEBUG nova.virt.hardware [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 801.146080] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created linked-clone VM from snapshot [ 801.147058] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70982d5-68df-4c12-a4b0-9c71628273fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.150920] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bad4480-1f2d-494f-83e4-68140ed95de7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.160532] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514757} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.161309] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk. [ 801.162074] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7e5554-afb0-404a-84c1-a1374e7dd5d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.173249] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5257d421-be9b-27fe-36b8-b4231305f678, 'name': SearchDatastore_Task, 'duration_secs': 0.008645} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.173381] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploading image 2acbb9d2-26b0-4d2e-bae5-6a49f851ff7e {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 801.176590] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e02e667-92e8-4ade-9fbb-1970b5fff5fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.194031] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db7524d7-7a96-4d9f-a935-f05e6169e69a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.204294] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.207574] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dabd276-ff92-4da6-abff-fdb39d5dde71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.235118] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 801.235118] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c899a0-3f1c-2674-5038-f074a3800125" [ 801.235118] env[61964]: _type = "Task" [ 801.235118] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.239367] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 801.239367] env[61964]: value = "vm-230473" [ 801.239367] env[61964]: _type = "VirtualMachine" [ 801.239367] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 801.240606] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-29fb0f4b-035d-4808-b63e-ff2352abdee2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.245225] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 801.245225] env[61964]: value = "task-1040947" [ 801.245225] env[61964]: _type = "Task" [ 801.245225] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.249234] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c899a0-3f1c-2674-5038-f074a3800125, 'name': SearchDatastore_Task, 'duration_secs': 0.009664} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.250617] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.250928] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 560bd013-eae1-4541-aaa2-0732d8f20e4a/560bd013-eae1-4541-aaa2-0732d8f20e4a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 801.251205] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease: (returnval){ [ 801.251205] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52af5051-fea9-ce91-76a6-337e7a8ccfd7" [ 801.251205] env[61964]: _type = "HttpNfcLease" [ 801.251205] env[61964]: } obtained for exporting VM: (result){ [ 801.251205] env[61964]: value = "vm-230473" [ 801.251205] env[61964]: _type = "VirtualMachine" [ 801.251205] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 801.251582] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the lease: (returnval){ [ 801.251582] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52af5051-fea9-ce91-76a6-337e7a8ccfd7" [ 801.251582] env[61964]: _type = "HttpNfcLease" [ 801.251582] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 801.251582] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8cff684c-7d99-4b68-b481-27b7210bc70c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.258858] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.265825] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 801.265825] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52af5051-fea9-ce91-76a6-337e7a8ccfd7" [ 801.265825] env[61964]: _type = "HttpNfcLease" [ 801.265825] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 801.267269] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 801.267269] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52af5051-fea9-ce91-76a6-337e7a8ccfd7" [ 801.267269] env[61964]: _type = "HttpNfcLease" [ 801.267269] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 801.267733] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 801.267733] env[61964]: value = "task-1040949" [ 801.267733] env[61964]: _type = "Task" [ 801.267733] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.271742] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c261279d-0633-4401-a98a-5ebd7a882117 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.280953] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 801.280953] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 801.281597] env[61964]: DEBUG nova.network.neutron [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Updating instance_info_cache with network_info: [{"id": "24e07767-885b-4cd5-9127-da8635ce0092", "address": "fa:16:3e:7d:59:f3", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e07767-88", "ovs_interfaceid": "24e07767-885b-4cd5-9127-da8635ce0092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.286222] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040949, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.380355] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f934cfa0-f7c6-4d55-acc4-19f725b64a1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.546859] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.547515] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 801.554993] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.626s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.554993] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.558181] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.094s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.560457] env[61964]: INFO nova.compute.claims [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.618560] env[61964]: INFO nova.scheduler.client.report [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Deleted allocations for instance 0156727e-60dc-4e24-99e6-96b3e735aa0e [ 801.627646] env[61964]: INFO nova.compute.manager [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Rebuilding instance [ 801.679753] env[61964]: DEBUG nova.compute.manager [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.680911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a43329b-3c4d-4bf0-b867-bb2f1cb1cda8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.757457] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040947, 'name': ReconfigVM_Task, 'duration_secs': 0.384659} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.757457] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfigured VM instance instance-0000003c to attach disk [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba/5f28385d-6ea0-420d-8a26-4cb693714c14-rescue.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.757457] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e265d5-d05c-4263-b784-fc196ba3f00d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.796035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.796035] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance network_info: |[{"id": "24e07767-885b-4cd5-9127-da8635ce0092", "address": "fa:16:3e:7d:59:f3", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e07767-88", "ovs_interfaceid": "24e07767-885b-4cd5-9127-da8635ce0092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 801.796035] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c3c7d33-89d6-421a-8a5e-a01c7e2e15d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.807079] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Acquired lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.807304] env[61964]: DEBUG nova.network.neutron [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Refreshing network info cache for port 24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.809409] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:59:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24e07767-885b-4cd5-9127-da8635ce0092', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.821344] env[61964]: DEBUG oslo.service.loopingcall [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.821801] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.822882] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6aa340ae-92f9-4801-922c-2d3c7f5e56d1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.846204] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 801.846204] env[61964]: value = "task-1040950" [ 801.846204] env[61964]: _type = "Task" [ 801.846204] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.847034] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040949, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511713} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.847560] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 560bd013-eae1-4541-aaa2-0732d8f20e4a/560bd013-eae1-4541-aaa2-0732d8f20e4a.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.847852] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.851079] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e609073-f055-4dbe-b0d2-b434421ac718 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.855297] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.855297] env[61964]: value = "task-1040951" [ 801.855297] env[61964]: _type = "Task" [ 801.855297] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.863415] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040950, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.864079] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 801.864079] env[61964]: value = "task-1040952" [ 801.864079] env[61964]: _type = "Task" [ 801.864079] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.871415] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040951, 'name': CreateVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.875952] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.021826] env[61964]: DEBUG nova.compute.manager [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received event network-vif-plugged-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 802.022154] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.022397] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.022576] env[61964]: DEBUG oslo_concurrency.lockutils [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.022787] env[61964]: DEBUG nova.compute.manager [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] No waiting events found dispatching network-vif-plugged-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 802.024187] env[61964]: WARNING nova.compute.manager [req-bf0d7492-7e9d-4f8d-8c00-60487c51b936 req-f028642a-7885-424c-90c7-c9068d6b96e2 service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received unexpected event network-vif-plugged-1c91d247-4361-482e-a473-d6cc5926a4ca for instance with vm_state building and task_state spawning. [ 802.068321] env[61964]: DEBUG nova.compute.utils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.073037] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 802.073037] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.134774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-23917f9a-1f0a-4825-a119-97e5a8280c72 tempest-ServersTestMultiNic-998292300 tempest-ServersTestMultiNic-998292300-project-member] Lock "0156727e-60dc-4e24-99e6-96b3e735aa0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.315s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.142797] env[61964]: DEBUG nova.policy [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.358019] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040950, 'name': ReconfigVM_Task, 'duration_secs': 0.318338} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.364275] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 802.364275] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9779863-0633-4ff5-ac23-865a66372f10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.372131] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040951, 'name': CreateVM_Task, 'duration_secs': 0.38447} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.374165] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.374541] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 802.374541] env[61964]: value = "task-1040953" [ 802.374541] env[61964]: _type = "Task" [ 802.374541] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.375545] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.375963] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.376475] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.379781] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b8b37e-a11d-43bc-8bbd-2f3fbbc0fbcb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.387288] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072796} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.387288] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.387838] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e8b716-5ca7-4121-8ebc-ac8c57ff63db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.400938] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 802.400938] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5228ff5e-aea4-e100-0a7b-f8468fd52c76" [ 802.400938] env[61964]: _type = "Task" [ 802.400938] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.401514] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040953, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.428380] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 560bd013-eae1-4541-aaa2-0732d8f20e4a/560bd013-eae1-4541-aaa2-0732d8f20e4a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.429953] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Successfully updated port: 1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.433841] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79167b53-cd9c-4970-94a5-2f5298b2b6fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.457675] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5228ff5e-aea4-e100-0a7b-f8468fd52c76, 'name': SearchDatastore_Task, 'duration_secs': 0.011838} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.458803] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.459143] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.459688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.459913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.460672] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.461623] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e299942-d89b-4056-ab01-2140e24183f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.468165] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 802.468165] env[61964]: value = "task-1040954" [ 802.468165] env[61964]: _type = "Task" [ 802.468165] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.472835] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.473182] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 802.478281] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb6ea6ac-f1ec-44b0-a044-7f5111dd7f35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.480250] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040954, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.483900] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 802.483900] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d2c457-3488-43c3-0b9f-959ebfb91f37" [ 802.483900] env[61964]: _type = "Task" [ 802.483900] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.493300] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d2c457-3488-43c3-0b9f-959ebfb91f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.567234] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.567594] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.576083] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 802.703279] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 802.703651] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d2a38b2-09bd-4d63-a425-68ef8f8fc623 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.711435] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 802.711435] env[61964]: value = "task-1040955" [ 802.711435] env[61964]: _type = "Task" [ 802.711435] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.724813] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.818161] env[61964]: DEBUG nova.network.neutron [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Updated VIF entry in instance network info cache for port 24e07767-885b-4cd5-9127-da8635ce0092. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 802.818585] env[61964]: DEBUG nova.network.neutron [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Updating instance_info_cache with network_info: [{"id": "24e07767-885b-4cd5-9127-da8635ce0092", "address": "fa:16:3e:7d:59:f3", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e07767-88", "ovs_interfaceid": "24e07767-885b-4cd5-9127-da8635ce0092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.823005] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Successfully created port: 9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.887480] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040953, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.957773] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.957996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.958613] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.983862] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040954, 'name': ReconfigVM_Task, 'duration_secs': 0.331027} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.984358] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 560bd013-eae1-4541-aaa2-0732d8f20e4a/560bd013-eae1-4541-aaa2-0732d8f20e4a.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.985141] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3f8522b-5041-437d-99eb-9afe167900cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.999449] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d2c457-3488-43c3-0b9f-959ebfb91f37, 'name': SearchDatastore_Task, 'duration_secs': 0.011829} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.001956] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 803.001956] env[61964]: value = "task-1040956" [ 803.001956] env[61964]: _type = "Task" [ 803.001956] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.002300] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aa80cfd-c5a9-4abf-b20a-e92cd7789f42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.013085] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 803.013085] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e14206-427a-40bf-a3af-1837d4953742" [ 803.013085] env[61964]: _type = "Task" [ 803.013085] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.017189] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040956, 'name': Rename_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.027617] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e14206-427a-40bf-a3af-1837d4953742, 'name': SearchDatastore_Task, 'duration_secs': 0.01078} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.031286] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.031789] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 803.032406] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b80d13ee-955d-4326-8234-2d4f615b1a23 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.039881] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 803.039881] env[61964]: value = "task-1040957" [ 803.039881] env[61964]: _type = "Task" [ 803.039881] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.049561] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.071750] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 803.076551] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e95296-63dc-4e61-9295-63e8ccee381a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.092311] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0acef9-2a3e-4d40-8c23-a4c6f6bd9103 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.142143] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05accc6f-d38e-4c6f-91f6-9f6ea22f97bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.150905] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0446f034-41c7-45b1-94e2-fc6762dfd7fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.167237] env[61964]: DEBUG nova.compute.provider_tree [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.222685] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040955, 'name': PowerOffVM_Task, 'duration_secs': 0.225352} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.223085] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 803.223964] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 803.224288] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb02be24-d7be-4d4b-abce-7d291f990ac9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.231729] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 803.231729] env[61964]: value = "task-1040958" [ 803.231729] env[61964]: _type = "Task" [ 803.231729] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.245711] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 803.246176] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 803.246814] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230369', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'name': 'volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cdfc0bf1-d603-4580-8527-8b06e5ae0799', 'attached_at': '', 'detached_at': '', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'serial': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 803.247723] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8479f69-ec86-42da-8e53-d383cabef14a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.272877] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2b4d9c-23c0-482d-b4ef-35e244d02bfa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.282052] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847e943f-4d01-4d95-84ae-3c5ff85e9865 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.303084] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35f375a-76a0-4565-b41d-d7aa3a7c3926 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.320507] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] The volume has not been displaced from its original location: [datastore1] volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f/volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 803.326046] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Reconfiguring VM instance instance-00000023 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 803.326797] env[61964]: DEBUG oslo_concurrency.lockutils [req-aeeadafd-6f55-4fbb-9592-15a12fdd2fb5 req-2ef5d60c-71ca-4bcf-9748-eaa5a1f15714 service nova] Releasing lock "refresh_cache-a23e7f40-1b3e-4c8f-b664-30f6314923a1" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.327252] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06f7bc77-ab30-4e76-b5bd-854e06cb7b6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.347851] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 803.347851] env[61964]: value = "task-1040959" [ 803.347851] env[61964]: _type = "Task" [ 803.347851] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.357496] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.388934] env[61964]: DEBUG oslo_vmware.api [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040953, 'name': PowerOnVM_Task, 'duration_secs': 0.714958} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.389271] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 803.392685] env[61964]: DEBUG nova.compute.manager [None req-7da4cf16-75e7-4ecc-9bec-fb7a9942361f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 803.393618] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d41d6fb-0edd-4f04-b07a-a9478580ada1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.516395] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040956, 'name': Rename_Task, 'duration_secs': 0.156445} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.516727] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.516994] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a96b0270-6289-451f-adb5-704be53f61df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.526385] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 803.526385] env[61964]: value = "task-1040960" [ 803.526385] env[61964]: _type = "Task" [ 803.526385] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.536577] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.549600] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504411} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.549885] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 803.550508] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.550629] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78d746c5-b906-4256-b516-8cdb3c6e75de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.557026] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 803.557026] env[61964]: value = "task-1040961" [ 803.557026] env[61964]: _type = "Task" [ 803.557026] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.561941] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.570631] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.591990] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 803.597426] env[61964]: DEBUG nova.compute.manager [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 803.597696] env[61964]: DEBUG nova.compute.manager [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing instance network info cache due to event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 803.597953] env[61964]: DEBUG oslo_concurrency.lockutils [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] Acquiring lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.603170] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.622703] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 803.623104] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.623313] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 803.623492] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.623639] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 803.623787] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 803.624043] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 803.624257] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 803.624494] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 803.624743] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 803.624978] env[61964]: DEBUG nova.virt.hardware [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 803.625949] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69aa6ca-81c6-44f8-92ff-731c0ee6973b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.635290] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a28e295-7eee-4217-8ae3-63c46d8f3671 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.674552] env[61964]: DEBUG nova.scheduler.client.report [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 803.863810] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040959, 'name': ReconfigVM_Task, 'duration_secs': 0.32663} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.864088] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Reconfigured VM instance instance-00000023 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 803.870618] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-48da76f4-42e8-46cd-9b04-a10aa09d6b97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.890582] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 803.890582] env[61964]: value = "task-1040962" [ 803.890582] env[61964]: _type = "Task" [ 803.890582] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.900156] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040962, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.969273] env[61964]: DEBUG nova.network.neutron [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [{"id": "1c91d247-4361-482e-a473-d6cc5926a4ca", "address": "fa:16:3e:27:99:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c91d247-43", "ovs_interfaceid": "1c91d247-4361-482e-a473-d6cc5926a4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.040974] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040960, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.067147] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127185} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.067595] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.068541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11894bc2-ce74-4b2c-85bc-ff18a4d85fce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.097201] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.097606] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3db2117-2a85-44c3-869a-697608b53046 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.122207] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 804.122207] env[61964]: value = "task-1040963" [ 804.122207] env[61964]: _type = "Task" [ 804.122207] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.131527] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040963, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.181698] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.182386] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 804.185907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.227s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.186123] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.188289] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.694s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.188510] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.190600] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.032s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.190805] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.196815] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.950s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.196815] env[61964]: DEBUG nova.objects.instance [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'resources' on Instance uuid 595c94f1-9dec-454d-b301-5656287f53de {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.224318] env[61964]: INFO nova.scheduler.client.report [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Deleted allocations for instance 6ee845c3-5f79-4704-8b7b-cd3770202647 [ 804.226241] env[61964]: INFO nova.scheduler.client.report [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Deleted allocations for instance 65566c20-6d69-471c-b098-3c30c01d9955 [ 804.242374] env[61964]: INFO nova.scheduler.client.report [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Deleted allocations for instance af413b1d-9e7e-43d0-a39a-fb9dda9cf281 [ 804.402297] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040962, 'name': ReconfigVM_Task, 'duration_secs': 0.185} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.402851] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230369', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'name': 'volume-c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cdfc0bf1-d603-4580-8527-8b06e5ae0799', 'attached_at': '', 'detached_at': '', 'volume_id': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f', 'serial': 'c397e4a2-9c55-49e6-9341-0fa2a060c38f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 804.403349] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 804.404384] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d2bc18-16a4-45d4-944e-27cccd6efb3a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.413132] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 804.413132] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5bff8d7b-9a8d-49ee-b8fa-bb46314b255f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.473396] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.473396] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance network_info: |[{"id": "1c91d247-4361-482e-a473-d6cc5926a4ca", "address": "fa:16:3e:27:99:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c91d247-43", "ovs_interfaceid": "1c91d247-4361-482e-a473-d6cc5926a4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 804.473684] env[61964]: DEBUG oslo_concurrency.lockutils [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] Acquired lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.473990] env[61964]: DEBUG nova.network.neutron [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 804.479019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:99:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c91d247-4361-482e-a473-d6cc5926a4ca', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.484042] env[61964]: DEBUG oslo.service.loopingcall [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.486838] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.487185] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9533f3c7-d1b1-4901-9b1d-412ad34f35e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.508818] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 804.509263] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 804.509511] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Deleting the datastore file [datastore1] cdfc0bf1-d603-4580-8527-8b06e5ae0799 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.510412] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d413e07a-94b1-4204-897b-c4a039b941f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.516142] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.516142] env[61964]: value = "task-1040965" [ 804.516142] env[61964]: _type = "Task" [ 804.516142] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.520552] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for the task: (returnval){ [ 804.520552] env[61964]: value = "task-1040966" [ 804.520552] env[61964]: _type = "Task" [ 804.520552] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.529747] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040965, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.538571] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.546148] env[61964]: DEBUG oslo_vmware.api [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040960, 'name': PowerOnVM_Task, 'duration_secs': 0.513581} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.546148] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 804.546148] env[61964]: INFO nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Took 8.71 seconds to spawn the instance on the hypervisor. [ 804.546148] env[61964]: DEBUG nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 804.546148] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6be195-2702-4d4f-8447-5f825d35d3de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.633904] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040963, 'name': ReconfigVM_Task, 'duration_secs': 0.467156} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.634560] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Reconfigured VM instance instance-0000003f to attach disk [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.636279] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e90f79d4-6cbb-4ff5-864e-1d03dac5de21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.641200] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 804.641200] env[61964]: value = "task-1040967" [ 804.641200] env[61964]: _type = "Task" [ 804.641200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.650813] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040967, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.705017] env[61964]: DEBUG nova.compute.utils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.706848] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 804.707019] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 804.743648] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e16e96ad-8400-40d5-9b97-5cdab75fb459 tempest-MigrationsAdminTest-465935768 tempest-MigrationsAdminTest-465935768-project-member] Lock "65566c20-6d69-471c-b098-3c30c01d9955" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.440s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.745498] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10378f64-cb16-4627-b8df-6e8a2168016a tempest-ServersTestBootFromVolume-2138366211 tempest-ServersTestBootFromVolume-2138366211-project-member] Lock "6ee845c3-5f79-4704-8b7b-cd3770202647" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.795s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.754349] env[61964]: DEBUG oslo_concurrency.lockutils [None req-da453d6c-80ad-494d-a41f-8e244a55bc30 tempest-ServersTestFqdnHostnames-1324313166 tempest-ServersTestFqdnHostnames-1324313166-project-member] Lock "af413b1d-9e7e-43d0-a39a-fb9dda9cf281" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.857s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.793956] env[61964]: DEBUG nova.policy [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '771d16e2520d4c2e8161ba3da91a03ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19bac13e505143b4ae5884c24c748b1a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.822772] env[61964]: INFO nova.compute.manager [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Unrescuing [ 804.824661] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.824661] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquired lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.824661] env[61964]: DEBUG nova.network.neutron [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.031658] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040965, 'name': CreateVM_Task, 'duration_secs': 0.445877} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.031860] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.032995] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.033262] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.033645] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.033943] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cf0825a-8d64-44c0-9fc3-5be61c3cdd15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.038815] env[61964]: DEBUG oslo_vmware.api [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Task: {'id': task-1040966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108349} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.039386] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.039536] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 805.039720] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 805.044645] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 805.044645] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52928947-26d6-7a51-28f7-5c9f1f5c3a46" [ 805.044645] env[61964]: _type = "Task" [ 805.044645] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.054135] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52928947-26d6-7a51-28f7-5c9f1f5c3a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.068250] env[61964]: INFO nova.compute.manager [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Took 46.65 seconds to build instance. [ 805.131732] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 805.133349] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6062c1c4-a268-4d63-bfa1-9c02a5e2ef63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.138904] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df0fe01-bb85-4677-8648-7619cd77f1e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.152998] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbba52c-8dda-4435-91da-3df0efe97c35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.171387] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1756112-6db9-44e0-8d72-5a48a83d5450 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.179317] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040967, 'name': Rename_Task, 'duration_secs': 0.166085} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.180024] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.193261] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44176d5c-a8a2-4e74-957d-867f9bbb3368 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.196561] env[61964]: ERROR nova.compute.manager [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Failed to detach volume c397e4a2-9c55-49e6-9341-0fa2a060c38f from /dev/sda: nova.exception.InstanceNotFound: Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 could not be found. [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Traceback (most recent call last): [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self.driver.rebuild(**kwargs) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise NotImplementedError() [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] NotImplementedError [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] During handling of the above exception, another exception occurred: [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Traceback (most recent call last): [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self.driver.detach_volume(context, old_connection_info, [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] return self._volumeops.detach_volume(connection_info, instance) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._detach_volume_vmdk(connection_info, instance) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] stable_ref.fetch_moref(session) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise exception.InstanceNotFound(instance_id=self._uuid) [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] nova.exception.InstanceNotFound: Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 could not be found. [ 805.196561] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.225845] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 805.229919] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60803c6c-4d74-4d58-bd09-23dfe46adc24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.234165] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 805.234165] env[61964]: value = "task-1040968" [ 805.234165] env[61964]: _type = "Task" [ 805.234165] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.241316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1d77f4-e986-4403-8ac5-e3e18c369201 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.248610] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.262093] env[61964]: DEBUG nova.compute.provider_tree [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.397670] env[61964]: DEBUG nova.compute.utils [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Build of instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 aborted: Failed to rebuild volume backed instance. {{(pid=61964) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.405595] env[61964]: ERROR nova.compute.manager [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 aborted: Failed to rebuild volume backed instance. [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Traceback (most recent call last): [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self.driver.rebuild(**kwargs) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise NotImplementedError() [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] NotImplementedError [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] During handling of the above exception, another exception occurred: [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Traceback (most recent call last): [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._detach_root_volume(context, instance, root_bdm) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] with excutils.save_and_reraise_exception(): [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self.force_reraise() [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise self.value [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self.driver.detach_volume(context, old_connection_info, [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] return self._volumeops.detach_volume(connection_info, instance) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._detach_volume_vmdk(connection_info, instance) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] stable_ref.fetch_moref(session) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise exception.InstanceNotFound(instance_id=self._uuid) [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] nova.exception.InstanceNotFound: Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 could not be found. [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] During handling of the above exception, another exception occurred: [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Traceback (most recent call last): [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 11307, in _error_out_instance_on_exception [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] yield [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 805.405595] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._do_rebuild_instance_with_claim( [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._do_rebuild_instance( [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._rebuild_default_impl(**kwargs) [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] self._rebuild_volume_backed_instance( [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] raise exception.BuildAbortException( [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] nova.exception.BuildAbortException: Build of instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 aborted: Failed to rebuild volume backed instance. [ 805.409155] env[61964]: ERROR nova.compute.manager [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] [ 805.571234] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52928947-26d6-7a51-28f7-5c9f1f5c3a46, 'name': SearchDatastore_Task, 'duration_secs': 0.010795} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.571234] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.571234] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.571234] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.571234] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.571234] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.575019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6c742723-501d-4171-8a1c-f1d4374abc34 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.868s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.575019] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4a49954-fbb2-43a8-9e96-8a884c9ad2a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.583183] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.583457] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 805.584281] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f70958e8-9da5-4d9a-a94f-2537ff363865 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.592212] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 805.592212] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fcf59d-906b-0d4f-a844-9a06b5e240cc" [ 805.592212] env[61964]: _type = "Task" [ 805.592212] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.606109] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fcf59d-906b-0d4f-a844-9a06b5e240cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009581} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.610109] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90c20ee0-0136-40dd-82e9-0bf09855fe4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.616855] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 805.616855] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e0e59-e8f3-5e4d-0d57-b3ce030e23c4" [ 805.616855] env[61964]: _type = "Task" [ 805.616855] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.626741] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e0e59-e8f3-5e4d-0d57-b3ce030e23c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.750867] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040968, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.767910] env[61964]: DEBUG nova.scheduler.client.report [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.890174] env[61964]: DEBUG nova.network.neutron [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updated VIF entry in instance network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 805.890731] env[61964]: DEBUG nova.network.neutron [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [{"id": "1c91d247-4361-482e-a473-d6cc5926a4ca", "address": "fa:16:3e:27:99:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c91d247-43", "ovs_interfaceid": "1c91d247-4361-482e-a473-d6cc5926a4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.944048] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Successfully created port: df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.030074] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Successfully updated port: 9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.093788] env[61964]: DEBUG nova.compute.manager [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Received event network-vif-plugged-9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 806.094776] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.095718] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.096089] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.097270] env[61964]: DEBUG nova.compute.manager [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] No waiting events found dispatching network-vif-plugged-9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 806.097270] env[61964]: WARNING nova.compute.manager [req-4d35a90e-7a39-41e5-bf55-010f9601443c req-07852e34-9a19-4570-948e-7491d9ec0891 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Received unexpected event network-vif-plugged-9d295b40-9665-4671-9ea8-108dbb9169d7 for instance with vm_state building and task_state spawning. [ 806.134323] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523e0e59-e8f3-5e4d-0d57-b3ce030e23c4, 'name': SearchDatastore_Task, 'duration_secs': 0.01239} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.134667] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.134938] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ab33fb75-4114-4a26-8692-befb407b7460/ab33fb75-4114-4a26-8692-befb407b7460.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 806.135234] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53c380ba-a854-4db9-b18a-3218bc9a76e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.142847] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 806.142847] env[61964]: value = "task-1040969" [ 806.142847] env[61964]: _type = "Task" [ 806.142847] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.151770] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.236628] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 806.247891] env[61964]: DEBUG oslo_vmware.api [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040968, 'name': PowerOnVM_Task, 'duration_secs': 0.52951} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.248945] env[61964]: DEBUG nova.network.neutron [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updating instance_info_cache with network_info: [{"id": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "address": "fa:16:3e:51:4c:fb", "network": {"id": "09b3e2c7-fbe1-4d7c-be07-13064e28982a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1870835802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "787322c793384f849d0a2acae27a2052", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "412cde91-d0f0-4193-b36b-d8b9d17384c6", "external-id": "nsx-vlan-transportzone-461", "segmentation_id": 461, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e60e0a5-9e", "ovs_interfaceid": "6e60e0a5-9e9f-43f3-8823-cc1d7c0be746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.250154] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.250376] env[61964]: INFO nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Took 7.80 seconds to spawn the instance on the hypervisor. [ 806.250561] env[61964]: DEBUG nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 806.252333] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659fdba2-a1e7-49fa-8c4d-2b4610459ed8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.266470] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 806.266703] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.266865] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 806.267060] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.267216] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 806.267398] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 806.267569] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 806.267826] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 806.267993] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 806.268556] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 806.268746] env[61964]: DEBUG nova.virt.hardware [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 806.269596] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b867df1-bec8-4d52-9499-20c6203ce607 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.274884] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.078s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.281077] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.320s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.282591] env[61964]: INFO nova.compute.claims [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.287939] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216c6127-231a-4331-ace9-25c4127cea60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.322309] env[61964]: INFO nova.scheduler.client.report [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted allocations for instance 595c94f1-9dec-454d-b301-5656287f53de [ 806.393825] env[61964]: DEBUG oslo_concurrency.lockutils [req-546b0322-425e-44e3-b59f-c82b65def437 req-f438988f-20a5-4319-882c-77e1099926fd service nova] Releasing lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.500018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "062da016-f4d6-424d-90e6-5903940e2daa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.500018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.500018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "062da016-f4d6-424d-90e6-5903940e2daa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.500018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.500018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.501852] env[61964]: INFO nova.compute.manager [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Terminating instance [ 806.533178] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.533178] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.533309] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.655882] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040969, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.757544] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Releasing lock "refresh_cache-beddafaf-8a91-4bfa-af3c-fa69977677ba" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.757544] env[61964]: DEBUG nova.objects.instance [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lazy-loading 'flavor' on Instance uuid beddafaf-8a91-4bfa-af3c-fa69977677ba {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.774127] env[61964]: INFO nova.compute.manager [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Took 44.57 seconds to build instance. [ 806.830313] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f81bb6-5690-4d34-9ca2-513b74a502fd tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "595c94f1-9dec-454d-b301-5656287f53de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.078s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.007860] env[61964]: DEBUG nova.compute.manager [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.008146] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.009627] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ebc533-c131-4ac0-b4bc-d2a5ef9d1770 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.023094] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.023213] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a9db6b0-ec87-450b-8420-2d22c3c1ad21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.030838] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 807.030838] env[61964]: value = "task-1040970" [ 807.030838] env[61964]: _type = "Task" [ 807.030838] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.043453] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040970, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.111718] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 807.158975] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.743069} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.161813] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ab33fb75-4114-4a26-8692-befb407b7460/ab33fb75-4114-4a26-8692-befb407b7460.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.162674] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.162991] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92577e07-12f2-4be6-8ea2-be63e7f65886 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.171342] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 807.171342] env[61964]: value = "task-1040971" [ 807.171342] env[61964]: _type = "Task" [ 807.171342] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.182769] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.245647] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.246018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.247147] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.247479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.247821] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.251903] env[61964]: INFO nova.compute.manager [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Terminating instance [ 807.265558] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34643d0-8924-425d-b8c9-e3229a1c2b97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.293349] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f325b05c-a993-43d0-b1a5-429e1aedd78f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.374s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.297321] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.299264] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76df762a-abd5-4d18-bfc8-fdeff462b846 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.310485] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 807.310485] env[61964]: value = "task-1040972" [ 807.310485] env[61964]: _type = "Task" [ 807.310485] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.322364] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.423809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.493781] env[61964]: DEBUG nova.network.neutron [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Updating instance_info_cache with network_info: [{"id": "9d295b40-9665-4671-9ea8-108dbb9169d7", "address": "fa:16:3e:65:2b:9d", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d295b40-96", "ovs_interfaceid": "9d295b40-9665-4671-9ea8-108dbb9169d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.546946] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040970, 'name': PowerOffVM_Task, 'duration_secs': 0.320408} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.552983] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.552983] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.552983] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd91eab2-77ac-4830-8dd4-9af9f48a515a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.625972] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 807.625972] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 807.625972] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleting the datastore file [datastore1] 062da016-f4d6-424d-90e6-5903940e2daa {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.625972] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f3fb958-afb1-4675-b131-1115616eb995 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.634359] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 807.634359] env[61964]: value = "task-1040974" [ 807.634359] env[61964]: _type = "Task" [ 807.634359] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.644297] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.687226] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082896} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.687226] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.687226] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc571a0a-f31f-4e11-9962-b91cb4775a3a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.713851] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] ab33fb75-4114-4a26-8692-befb407b7460/ab33fb75-4114-4a26-8692-befb407b7460.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.716724] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dd43a4d-c997-40f9-accb-6233ce173502 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.740655] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 807.740655] env[61964]: value = "task-1040975" [ 807.740655] env[61964]: _type = "Task" [ 807.740655] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.746518] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bf55ee-6b7a-429b-8188-175e81c1420c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.753651] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040975, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.759555] env[61964]: DEBUG nova.compute.manager [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.762445] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.762445] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89b624d-8ca6-42e1-a8c4-2b06a3d9d54f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.766450] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c000787-0a07-4180-abdb-6f65f65586f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.777768] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.805038] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50933b67-6a07-4ebe-b839-66ff3bd2a23f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.807655] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fd37e3-2b3d-4ebb-a4ae-43a36683e8be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.817811] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 807.817811] env[61964]: value = "task-1040976" [ 807.817811] env[61964]: _type = "Task" [ 807.817811] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.819755] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e998fe31-35c4-482b-bb31-a71bf45e4881 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.831327] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040972, 'name': PowerOffVM_Task, 'duration_secs': 0.263659} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.832262] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.838413] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfiguring VM instance instance-0000003c to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 807.849743] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e73a1dfd-62f0-4f69-aefb-52f485225308 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.863381] env[61964]: DEBUG nova.compute.provider_tree [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.868372] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.874606] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 807.874606] env[61964]: value = "task-1040977" [ 807.874606] env[61964]: _type = "Task" [ 807.874606] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.885413] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.996606] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.997548] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Instance network_info: |[{"id": "9d295b40-9665-4671-9ea8-108dbb9169d7", "address": "fa:16:3e:65:2b:9d", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d295b40-96", "ovs_interfaceid": "9d295b40-9665-4671-9ea8-108dbb9169d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 807.997548] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:2b:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d295b40-9665-4671-9ea8-108dbb9169d7', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.006253] env[61964]: DEBUG oslo.service.loopingcall [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.006936] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 808.009019] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1add6088-75eb-4ab4-9346-7d7dddf480c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.033542] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.033542] env[61964]: value = "task-1040978" [ 808.033542] env[61964]: _type = "Task" [ 808.033542] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.045118] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040978, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.146952] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.255530] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040975, 'name': ReconfigVM_Task, 'duration_secs': 0.416962} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.256150] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Reconfigured VM instance instance-00000040 to attach disk [datastore2] ab33fb75-4114-4a26-8692-befb407b7460/ab33fb75-4114-4a26-8692-befb407b7460.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 808.256989] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b7d4a91-7da4-44ec-889d-38241d436195 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.266201] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 808.266201] env[61964]: value = "task-1040979" [ 808.266201] env[61964]: _type = "Task" [ 808.266201] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.280214] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040979, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.337732] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040976, 'name': PowerOffVM_Task, 'duration_secs': 0.23366} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.338410] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 808.338410] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 808.338670] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23fcc77e-64be-46a1-8e15-d555e0633347 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.370213] env[61964]: DEBUG nova.scheduler.client.report [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 808.386434] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.386824] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.394185] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040977, 'name': ReconfigVM_Task, 'duration_secs': 0.384966} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.394185] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Reconfigured VM instance instance-0000003c to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 808.394185] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.394185] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbb6b56f-8734-4084-b21e-3e7f0fd6fd42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.401222] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 808.401222] env[61964]: value = "task-1040981" [ 808.401222] env[61964]: _type = "Task" [ 808.401222] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.415522] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040981, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.502453] env[61964]: DEBUG nova.compute.manager [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Received event network-changed-9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 808.502453] env[61964]: DEBUG nova.compute.manager [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Refreshing instance network info cache due to event network-changed-9d295b40-9665-4671-9ea8-108dbb9169d7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 808.502453] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] Acquiring lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.503171] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] Acquired lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.503444] env[61964]: DEBUG nova.network.neutron [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Refreshing network info cache for port 9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.534152] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Successfully updated port: df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.546071] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040978, 'name': CreateVM_Task, 'duration_secs': 0.371016} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.550781] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.551541] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.551713] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.552067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.552628] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d93fa74-29cb-47a5-a90c-3f546a6e9750 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.567027] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 808.567027] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52319acd-5ab3-8bb7-e698-c992884fb302" [ 808.567027] env[61964]: _type = "Task" [ 808.567027] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.583021] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52319acd-5ab3-8bb7-e698-c992884fb302, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.645879] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.778392] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040979, 'name': Rename_Task, 'duration_secs': 0.172788} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.778845] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 808.779180] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cddc1241-ebb0-4a78-949a-9ef871b133f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.786647] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 808.786647] env[61964]: value = "task-1040982" [ 808.786647] env[61964]: _type = "Task" [ 808.786647] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.797420] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.875484] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.876993] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 808.880543] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.794s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.881144] env[61964]: DEBUG nova.objects.instance [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'resources' on Instance uuid 9adedaa2-8594-4ddd-8f85-a41174207ef8 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.894155] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 808.918668] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040981, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.038540] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.038540] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.038540] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.075847] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52319acd-5ab3-8bb7-e698-c992884fb302, 'name': SearchDatastore_Task, 'duration_secs': 0.028356} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.076290] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.076744] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.076836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.077029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.077193] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.077888] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e42a097-54e0-4397-a736-4c0c8068a7c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.087014] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.087283] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.087995] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff877c80-a8b4-4d42-b6be-09eefa066aa3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.094329] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 809.094329] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba7daf-3afa-10b9-ccdb-a119af64e802" [ 809.094329] env[61964]: _type = "Task" [ 809.094329] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.103477] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba7daf-3afa-10b9-ccdb-a119af64e802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.145417] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.309901] env[61964]: DEBUG oslo_vmware.api [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1040982, 'name': PowerOnVM_Task, 'duration_secs': 0.51551} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.310245] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 809.310370] env[61964]: INFO nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 8.21 seconds to spawn the instance on the hypervisor. [ 809.311021] env[61964]: DEBUG nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 809.313154] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed9ac73-4ce0-4179-bc9e-7f1fcf222f40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.385290] env[61964]: DEBUG nova.compute.utils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.387796] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 809.388048] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.399025] env[61964]: DEBUG nova.network.neutron [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Updated VIF entry in instance network info cache for port 9d295b40-9665-4671-9ea8-108dbb9169d7. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.399478] env[61964]: DEBUG nova.network.neutron [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Updating instance_info_cache with network_info: [{"id": "9d295b40-9665-4671-9ea8-108dbb9169d7", "address": "fa:16:3e:65:2b:9d", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d295b40-96", "ovs_interfaceid": "9d295b40-9665-4671-9ea8-108dbb9169d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.422348] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.422625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.422830] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.423482] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.423764] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.425785] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040981, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.426287] env[61964]: INFO nova.compute.manager [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Terminating instance [ 809.433180] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 809.433180] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 809.433180] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleting the datastore file [datastore1] 560bd013-eae1-4541-aaa2-0732d8f20e4a {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.434677] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.434923] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ac0e2d64-3343-438c-9f72-238cee9cdcc8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.440840] env[61964]: DEBUG nova.policy [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05bfff4e5dae488593812635d1c216b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f1804a0ded74317a1e2016db18c55ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.443539] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for the task: (returnval){ [ 809.443539] env[61964]: value = "task-1040983" [ 809.443539] env[61964]: _type = "Task" [ 809.443539] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.456497] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.608149] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba7daf-3afa-10b9-ccdb-a119af64e802, 'name': SearchDatastore_Task, 'duration_secs': 0.012807} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.608149] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.610685] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d35d158-3aba-4749-b0e7-31892aa3bd5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.617177] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 809.617177] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52092ccc-abb5-08be-4207-b17fec2dacad" [ 809.617177] env[61964]: _type = "Task" [ 809.617177] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.629783] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52092ccc-abb5-08be-4207-b17fec2dacad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.652721] env[61964]: DEBUG oslo_vmware.api [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040974, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.740955} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.653160] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.653398] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.653581] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.653759] env[61964]: INFO nova.compute.manager [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Took 2.65 seconds to destroy the instance on the hypervisor. [ 809.654188] env[61964]: DEBUG oslo.service.loopingcall [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.654411] env[61964]: DEBUG nova.compute.manager [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.654501] env[61964]: DEBUG nova.network.neutron [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 809.774755] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd37b67-e9c3-49f7-a238-a4c752e7a749 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.783530] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c5ea07-182c-47bd-a79c-03f119eb3539 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.829740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5220a78-1dc8-4725-bcf4-ff6082db513b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.849602] env[61964]: INFO nova.compute.manager [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 42.67 seconds to build instance. [ 809.849602] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6f52b6-039b-4344-aef9-5a852b6a0c87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.864233] env[61964]: DEBUG nova.compute.provider_tree [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.894361] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 809.907697] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4b1983-d3e5-4a62-a8a7-0c21d453c881 req-a7bbf2ab-a1e9-490f-950c-30be42c1ae88 service nova] Releasing lock "refresh_cache-1c922228-4ba9-43ed-8e5c-3ca15e48ceab" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.914839] env[61964]: DEBUG oslo_vmware.api [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040981, 'name': PowerOnVM_Task, 'duration_secs': 1.453254} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.915185] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 809.915491] env[61964]: DEBUG nova.compute.manager [None req-e54e85cf-d11d-4ab9-b04c-5c973eccc0cc tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 809.916343] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ba8fdc-72a7-4089-b5ab-de1e816d2eec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.940359] env[61964]: DEBUG nova.compute.manager [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 809.942235] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18d9d2ca-77eb-4196-bccc-69febccdd0a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.960100] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f51fcc4-28a2-439b-a84a-e5c266a79f0b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.975689] env[61964]: DEBUG oslo_vmware.api [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Task: {'id': task-1040983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253524} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.977898] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.978235] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 809.978477] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 809.978756] env[61964]: INFO nova.compute.manager [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Took 2.22 seconds to destroy the instance on the hypervisor. [ 809.979267] env[61964]: DEBUG oslo.service.loopingcall [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.981026] env[61964]: DEBUG nova.compute.manager [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.981026] env[61964]: DEBUG nova.network.neutron [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.004752] env[61964]: WARNING nova.virt.vmwareapi.driver [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 could not be found. [ 810.004752] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 810.004752] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbe2a861-36e1-491b-b59d-6027fd22dc58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.014317] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bd60d5-97a6-4b87-8cc2-f8478dc2e088 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.026810] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Successfully created port: e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.051084] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cdfc0bf1-d603-4580-8527-8b06e5ae0799 could not be found. [ 810.051788] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 810.051788] env[61964]: INFO nova.compute.manager [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Took 0.11 seconds to destroy the instance on the hypervisor. [ 810.052083] env[61964]: DEBUG oslo.service.loopingcall [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.052613] env[61964]: DEBUG nova.compute.manager [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 810.052730] env[61964]: DEBUG nova.network.neutron [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.136635] env[61964]: DEBUG nova.network.neutron [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.140130] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52092ccc-abb5-08be-4207-b17fec2dacad, 'name': SearchDatastore_Task, 'duration_secs': 0.013051} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.140130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.140130] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 1c922228-4ba9-43ed-8e5c-3ca15e48ceab/1c922228-4ba9-43ed-8e5c-3ca15e48ceab.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.140130] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2ecc0fa-c7cc-4b06-8b0b-1144db77c116 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.146939] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 810.146939] env[61964]: value = "task-1040984" [ 810.146939] env[61964]: _type = "Task" [ 810.146939] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.157200] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.359026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-700f28b4-7695-4fdc-bce9-2b4141406f87 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.968s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.367622] env[61964]: DEBUG nova.scheduler.client.report [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.643368] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.643600] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Instance network_info: |[{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 810.644205] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:c8:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '418ddd3d-5f64-407e-8e0c-c8b81639bee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df79c16f-2899-499e-ab0f-f8f7752913e5', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.652521] env[61964]: DEBUG oslo.service.loopingcall [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.652912] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 810.656759] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e81e5bb7-2282-422a-ba09-1e2fa87c31f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.683311] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040984, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.684994] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.684994] env[61964]: value = "task-1040985" [ 810.684994] env[61964]: _type = "Task" [ 810.684994] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.694097] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040985, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.775868] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "93720ce5-834a-4d8d-b230-13d38620b688" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.776384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.874450] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.994s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.878120] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.499s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.878120] env[61964]: DEBUG nova.objects.instance [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lazy-loading 'resources' on Instance uuid 4104dee9-34db-4396-9189-8dc677d4b677 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.906871] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 810.910592] env[61964]: INFO nova.scheduler.client.report [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted allocations for instance 9adedaa2-8594-4ddd-8f85-a41174207ef8 [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 810.953023] env[61964]: DEBUG nova.virt.hardware [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 810.953023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089bffa2-571b-4a4e-bccd-7ccee802329a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.960730] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac973bbc-aa6f-4c3c-8878-67fc76515a49 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.166052] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60406} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.167273] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 1c922228-4ba9-43ed-8e5c-3ca15e48ceab/1c922228-4ba9-43ed-8e5c-3ca15e48ceab.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.167846] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.168360] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc5cdff7-d939-49c1-b5b1-de7cf30825a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.176686] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 811.176686] env[61964]: value = "task-1040986" [ 811.176686] env[61964]: _type = "Task" [ 811.176686] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.193267] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.200945] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040985, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.268198] env[61964]: DEBUG nova.network.neutron [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.281696] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 811.419129] env[61964]: DEBUG oslo_concurrency.lockutils [None req-97044966-bbd0-4c29-8fa0-fea0424e768a tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "9adedaa2-8594-4ddd-8f85-a41174207ef8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.279s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.482780] env[61964]: DEBUG nova.network.neutron [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.634720] env[61964]: INFO nova.compute.manager [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Rebuilding instance [ 811.687671] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125899} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.696463] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.701167] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ca4943-2ec8-419e-952f-08f58d34995d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.702686] env[61964]: DEBUG nova.compute.manager [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.703529] env[61964]: DEBUG nova.network.neutron [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.705064] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c277f1-ebdf-44d0-a57f-501359f93d5b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.711762] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1040985, 'name': CreateVM_Task, 'duration_secs': 0.566377} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.714807] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.726139] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.726139] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.726312] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.736791] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 1c922228-4ba9-43ed-8e5c-3ca15e48ceab/1c922228-4ba9-43ed-8e5c-3ca15e48ceab.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.740808] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5ef46ac-5286-4e65-a6b6-ac266fe69c42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.742691] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7aeb7051-42d0-4ae9-b1bc-cc7893b4e4db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.761232] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354b0c01-e898-4adb-b325-e17fe8337a6e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.769043] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 811.769043] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ac6f-438f-b897-e636-5f01ac58dc8a" [ 811.769043] env[61964]: _type = "Task" [ 811.769043] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.771300] env[61964]: INFO nova.compute.manager [-] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Took 2.12 seconds to deallocate network for instance. [ 811.779542] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e2978d-b612-4c68-918b-8c7f0c2b0d50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.781743] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 811.781743] env[61964]: value = "task-1040987" [ 811.781743] env[61964]: _type = "Task" [ 811.781743] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.826576] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ac6f-438f-b897-e636-5f01ac58dc8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.830644] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e46ee36-a45d-47ac-96bb-4da961dd2cd7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.839063] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.845592] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfaee48-f4a9-4852-8967-b3616927a7c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.854070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.864014] env[61964]: DEBUG nova.compute.provider_tree [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.985176] env[61964]: INFO nova.compute.manager [-] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Took 2.00 seconds to deallocate network for instance. [ 812.004914] env[61964]: DEBUG nova.compute.manager [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received event network-vif-plugged-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 812.006129] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Acquiring lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.006129] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.006129] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.006129] env[61964]: DEBUG nova.compute.manager [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] No waiting events found dispatching network-vif-plugged-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 812.006310] env[61964]: WARNING nova.compute.manager [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received unexpected event network-vif-plugged-df79c16f-2899-499e-ab0f-f8f7752913e5 for instance with vm_state building and task_state spawning. [ 812.006453] env[61964]: DEBUG nova.compute.manager [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 812.006646] env[61964]: DEBUG nova.compute.manager [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing instance network info cache due to event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 812.006888] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Acquiring lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.007110] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Acquired lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.007327] env[61964]: DEBUG nova.network.neutron [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 812.133309] env[61964]: DEBUG nova.compute.manager [req-44b5a6da-6772-4105-9628-ce709e2d633e req-defcac07-b5d8-463f-9072-1dccb6aad59f service nova] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Received event network-vif-deleted-7210f188-7099-4e52-ad5f-f20da7bc5d91 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 812.213033] env[61964]: INFO nova.compute.manager [-] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Took 2.16 seconds to deallocate network for instance. [ 812.288129] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ac6f-438f-b897-e636-5f01ac58dc8a, 'name': SearchDatastore_Task, 'duration_secs': 0.026878} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.292447] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.292447] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.292606] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.292735] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.292925] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.294117] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-996f828a-1b74-42e7-bb88-b04b30739fc4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.301210] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040987, 'name': ReconfigVM_Task, 'duration_secs': 0.418764} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.303013] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 1c922228-4ba9-43ed-8e5c-3ca15e48ceab/1c922228-4ba9-43ed-8e5c-3ca15e48ceab.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.306930] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a522132-b90e-428a-809b-8cf6bc020160 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.309272] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.309272] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 812.310254] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6f613eb-2225-4893-b2fc-5fe8e4f75384 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.321052] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 812.321052] env[61964]: value = "task-1040988" [ 812.321052] env[61964]: _type = "Task" [ 812.321052] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.323922] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 812.323922] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b36c13-9960-81f1-ba76-6aece010325e" [ 812.323922] env[61964]: _type = "Task" [ 812.323922] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.324884] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.340680] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b36c13-9960-81f1-ba76-6aece010325e, 'name': SearchDatastore_Task, 'duration_secs': 0.017956} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.354143] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040988, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.356298] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac50e59c-f690-4a1b-bbd9-6943e228426f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.363808] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 812.363808] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c9377c-bc1e-06db-166b-5146206e8fcf" [ 812.363808] env[61964]: _type = "Task" [ 812.363808] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.370037] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 812.371087] env[61964]: DEBUG nova.scheduler.client.report [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 812.378039] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bb83c4-8fb1-420c-b638-af6a3c09c86e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.384927] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c9377c-bc1e-06db-166b-5146206e8fcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.387690] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 812.387690] env[61964]: ERROR oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk due to incomplete transfer. [ 812.387690] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f1f6b85a-b23b-4e32-ae12-ed268395bf7f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.396392] env[61964]: DEBUG oslo_vmware.rw_handles [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d0a8a9-47b2-6dd9-ced4-7144be729208/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 812.396583] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploaded image 2acbb9d2-26b0-4d2e-bae5-6a49f851ff7e to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 812.399107] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 812.399864] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-da5c0180-5381-48c2-98c4-b0fffbf0ffde {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.406365] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 812.406365] env[61964]: value = "task-1040989" [ 812.406365] env[61964]: _type = "Task" [ 812.406365] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.416546] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040989, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.512029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.771058] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 812.771418] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1946093f-7953-4efb-b2c4-296c44ac698d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.777562] env[61964]: INFO nova.compute.manager [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Took 0.57 seconds to detach 1 volumes for instance. [ 812.780825] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 812.780825] env[61964]: value = "task-1040990" [ 812.780825] env[61964]: _type = "Task" [ 812.780825] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.781374] env[61964]: DEBUG nova.compute.manager [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Deleting volume: c397e4a2-9c55-49e6-9341-0fa2a060c38f {{(pid=61964) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 812.797541] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.811552] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.811930] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.812189] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.812425] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.812711] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.815905] env[61964]: INFO nova.compute.manager [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Terminating instance [ 812.843838] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040988, 'name': Rename_Task, 'duration_secs': 0.475955} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.844165] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.844432] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6461f8e-0767-48f1-a1a4-7be03e3089ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.851503] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 812.851503] env[61964]: value = "task-1040991" [ 812.851503] env[61964]: _type = "Task" [ 812.851503] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.859964] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040991, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.874208] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c9377c-bc1e-06db-166b-5146206e8fcf, 'name': SearchDatastore_Task, 'duration_secs': 0.022354} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.874544] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.874755] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 2a185e4b-c636-418a-a75e-2cf158f550c0/2a185e4b-c636-418a-a75e-2cf158f550c0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.875054] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a5f86c2-0b50-40ae-ac53-5efa46a5db61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.878811] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.882957] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.911s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.883024] env[61964]: DEBUG nova.objects.instance [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.886551] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 812.886551] env[61964]: value = "task-1040993" [ 812.886551] env[61964]: _type = "Task" [ 812.886551] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.897682] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Successfully updated port: e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.903281] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.918674] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040989, 'name': Destroy_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.931788] env[61964]: INFO nova.scheduler.client.report [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Deleted allocations for instance 4104dee9-34db-4396-9189-8dc677d4b677 [ 813.159089] env[61964]: DEBUG nova.network.neutron [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updated VIF entry in instance network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.159089] env[61964]: DEBUG nova.network.neutron [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.295383] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040990, 'name': PowerOffVM_Task, 'duration_secs': 0.221451} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.295383] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 813.295383] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.296304] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496ab750-110c-4d02-b652-b9b8f289a4bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.305164] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 813.305164] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d37aa71-3733-4ef2-bb71-697ad1c761e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.320605] env[61964]: DEBUG nova.compute.manager [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 813.324017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.324017] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc4213-f64a-4229-86c2-ef89aa53f81e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.332127] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.332127] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-993429fe-1364-4e2c-8cc4-e6be73832304 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.339157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.341318] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 813.341318] env[61964]: value = "task-1040995" [ 813.341318] env[61964]: _type = "Task" [ 813.341318] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.354985] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.365094] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040991, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.377960] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 813.378647] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 813.378980] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.379405] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aec84cd1-7f06-4caa-8c06-e67dc8be3de4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.388498] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 813.388498] env[61964]: value = "task-1040996" [ 813.388498] env[61964]: _type = "Task" [ 813.388498] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.401256] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.407689] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.407913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.411052] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.411052] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515369} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.415716] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 2a185e4b-c636-418a-a75e-2cf158f550c0/2a185e4b-c636-418a-a75e-2cf158f550c0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.415716] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.415716] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81d8883d-8818-4d5b-90c0-e04a73cb106b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.426048] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040989, 'name': Destroy_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.430480] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 813.430480] env[61964]: value = "task-1040997" [ 813.430480] env[61964]: _type = "Task" [ 813.430480] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.441877] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040997, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.442362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-344ab40f-bab1-43ad-9373-f493e2e5e2fe tempest-ServersV294TestFqdnHostnames-1635348598 tempest-ServersV294TestFqdnHostnames-1635348598-project-member] Lock "4104dee9-34db-4396-9189-8dc677d4b677" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.583s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.662723] env[61964]: DEBUG oslo_concurrency.lockutils [req-90ed51ee-2c3c-470e-8f84-e25e218e8442 req-fabe0da1-228e-4409-a740-f050dcfa9dd2 service nova] Releasing lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.839409] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586ec818-9a40-4850-b892-9b253c4f4bef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.856618] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55d11c3-e1db-438f-8f19-af897d96eb59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.860100] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.868357] env[61964]: DEBUG oslo_vmware.api [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1040991, 'name': PowerOnVM_Task, 'duration_secs': 0.934246} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.899106] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 813.899751] env[61964]: INFO nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Took 10.31 seconds to spawn the instance on the hypervisor. [ 813.899751] env[61964]: DEBUG nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 813.900811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d611308-1b55-47ab-905d-745e63f6efa8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.907071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca7a201-c804-4fb5-93b5-d0d9ccad4efa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.922518] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1040996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145213} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.927740] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.927740] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 813.927740] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 813.929763] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56cbbe4d-b9a3-4824-ae2f-8526cc14ee38 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.940299] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040989, 'name': Destroy_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.956740] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040997, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132768} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.956808] env[61964]: DEBUG nova.compute.provider_tree [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.958090] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.959041] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24307862-a48a-43af-acdc-3062089a7376 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.983717] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 2a185e4b-c636-418a-a75e-2cf158f550c0/2a185e4b-c636-418a-a75e-2cf158f550c0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.984720] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.988083] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-116145f0-b953-4464-9b52-628154ef8218 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.011879] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 814.011879] env[61964]: value = "task-1040998" [ 814.011879] env[61964]: _type = "Task" [ 814.011879] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.020465] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040998, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.228525] env[61964]: DEBUG nova.network.neutron [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Updating instance_info_cache with network_info: [{"id": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "address": "fa:16:3e:c9:8d:e0", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2ef8b26-3c", "ovs_interfaceid": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.364179] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.424053] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040989, 'name': Destroy_Task, 'duration_secs': 1.836194} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.425485] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroyed the VM [ 814.425945] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 814.426626] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-81201232-9f9f-4af6-bd43-74b8d34879a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.443322] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 814.443322] env[61964]: value = "task-1040999" [ 814.443322] env[61964]: _type = "Task" [ 814.443322] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.450498] env[61964]: INFO nova.compute.manager [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Took 46.73 seconds to build instance. [ 814.463023] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040999, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.463023] env[61964]: DEBUG nova.scheduler.client.report [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.522807] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1040998, 'name': ReconfigVM_Task, 'duration_secs': 0.291676} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.523207] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 2a185e4b-c636-418a-a75e-2cf158f550c0/2a185e4b-c636-418a-a75e-2cf158f550c0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.523922] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66b793e2-35b2-4b47-80ac-d3e8ce1a6f79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.531024] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 814.531024] env[61964]: value = "task-1041000" [ 814.531024] env[61964]: _type = "Task" [ 814.531024] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.540598] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041000, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.737182] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.737182] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Instance network_info: |[{"id": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "address": "fa:16:3e:c9:8d:e0", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2ef8b26-3c", "ovs_interfaceid": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 814.737182] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:8d:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2ef8b26-3c97-46b6-b6d2-407abf5737c9', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.745385] env[61964]: DEBUG oslo.service.loopingcall [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.745669] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 814.745901] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0995f843-5c43-447b-bdf7-8541305c914c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.771566] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 814.771566] env[61964]: value = "task-1041001" [ 814.771566] env[61964]: _type = "Task" [ 814.771566] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.780763] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041001, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.861735] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1040995, 'name': PowerOffVM_Task, 'duration_secs': 1.23614} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.862253] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.862504] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.862833] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79f06ad0-b06e-44d2-9470-41a2af73c8f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.926356] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.928284] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.928284] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleting the datastore file [datastore1] beddafaf-8a91-4bfa-af3c-fa69977677ba {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.928284] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04e9656d-06b6-4cb8-84ad-1f1503d8bd5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.933430] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 814.933430] env[61964]: value = "task-1041003" [ 814.933430] env[61964]: _type = "Task" [ 814.933430] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.950093] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.952849] env[61964]: DEBUG oslo_concurrency.lockutils [None req-008a49cf-6b25-4128-831c-3f3cdb021246 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.136s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.964439] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040999, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.969563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.974468] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.280s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.977825] env[61964]: INFO nova.compute.claims [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.998936] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 815.002806] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 815.003421] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 815.003701] env[61964]: DEBUG nova.virt.hardware [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 815.004985] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfa1098-e549-4754-ae98-ee2ef615412e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.015650] env[61964]: DEBUG nova.compute.manager [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Received event network-vif-deleted-dc3f581d-2c02-4c24-a56e-e9e2923931cb {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 815.015955] env[61964]: DEBUG nova.compute.manager [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Received event network-vif-deleted-7ac45fc2-5aeb-4136-85ad-abb23c24b047 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 815.016306] env[61964]: DEBUG nova.compute.manager [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 815.016548] env[61964]: DEBUG nova.compute.manager [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing instance network info cache due to event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 815.016871] env[61964]: DEBUG oslo_concurrency.lockutils [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.017087] env[61964]: DEBUG oslo_concurrency.lockutils [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.017311] env[61964]: DEBUG nova.network.neutron [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 815.023828] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0259771b-9a19-4d16-ac8a-9929b156d391 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.040686] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:59:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24e07767-885b-4cd5-9127-da8635ce0092', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.049784] env[61964]: DEBUG oslo.service.loopingcall [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.051736] env[61964]: INFO nova.scheduler.client.report [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d [ 815.060844] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 815.066548] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Received event network-vif-plugged-e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 815.066787] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquiring lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.067009] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.067185] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.067573] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] No waiting events found dispatching network-vif-plugged-e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 815.067799] env[61964]: WARNING nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Received unexpected event network-vif-plugged-e2ef8b26-3c97-46b6-b6d2-407abf5737c9 for instance with vm_state building and task_state spawning. [ 815.068493] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Received event network-changed-e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 815.068493] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Refreshing instance network info cache due to event network-changed-e2ef8b26-3c97-46b6-b6d2-407abf5737c9. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 815.068493] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquiring lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.068638] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquired lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.068860] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Refreshing network info cache for port e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 815.071356] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-986a5f0b-3afe-4088-a8f3-c59fd884cb1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.099761] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041000, 'name': Rename_Task, 'duration_secs': 0.142744} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.101335] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 815.101538] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.101538] env[61964]: value = "task-1041004" [ 815.101538] env[61964]: _type = "Task" [ 815.101538] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.102015] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5987a81-22a6-41df-95ab-eab8cf5bf31e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.113615] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041004, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.115393] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 815.115393] env[61964]: value = "task-1041005" [ 815.115393] env[61964]: _type = "Task" [ 815.115393] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.126474] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.291563] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041001, 'name': CreateVM_Task, 'duration_secs': 0.419014} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.291563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 815.291563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.291563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.292140] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.292140] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-444fd85c-612d-45d9-8309-81222360b71c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.297017] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 815.297017] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525cf920-4795-250e-a9f1-cecd795033ae" [ 815.297017] env[61964]: _type = "Task" [ 815.297017] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.305363] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525cf920-4795-250e-a9f1-cecd795033ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.444751] env[61964]: DEBUG oslo_vmware.api [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16502} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.444822] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.444961] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.445185] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.445374] env[61964]: INFO nova.compute.manager [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Took 2.12 seconds to destroy the instance on the hypervisor. [ 815.445667] env[61964]: DEBUG oslo.service.loopingcall [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.445904] env[61964]: DEBUG nova.compute.manager [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 815.445982] env[61964]: DEBUG nova.network.neutron [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.455576] env[61964]: DEBUG oslo_vmware.api [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1040999, 'name': RemoveSnapshot_Task, 'duration_secs': 0.695443} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.456016] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 815.456128] env[61964]: INFO nova.compute.manager [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 18.01 seconds to snapshot the instance on the hypervisor. [ 815.592800] env[61964]: DEBUG oslo_concurrency.lockutils [None req-545ede7a-eed3-4a4e-b617-b3b294e981bd tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "6a03c9cc-bf79-4d60-98fd-48b2e5bc547d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.380s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.623496] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041004, 'name': CreateVM_Task, 'duration_secs': 0.419318} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.624579] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 815.624929] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.629272] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041005, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.680886] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Updated VIF entry in instance network info cache for port e2ef8b26-3c97-46b6-b6d2-407abf5737c9. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.681271] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Updating instance_info_cache with network_info: [{"id": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "address": "fa:16:3e:c9:8d:e0", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2ef8b26-3c", "ovs_interfaceid": "e2ef8b26-3c97-46b6-b6d2-407abf5737c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.810825] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525cf920-4795-250e-a9f1-cecd795033ae, 'name': SearchDatastore_Task, 'duration_secs': 0.012527} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.811392] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.811736] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.812176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.812425] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.812813] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.813435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.813589] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.813906] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93b9d487-8841-4e4a-adb2-5a9a64c01e4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.816800] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df36beaa-125b-401a-bae8-2eec47cc0292 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.823139] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 815.823139] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520eb264-db56-7835-27a2-634f0e6418d1" [ 815.823139] env[61964]: _type = "Task" [ 815.823139] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.842168] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520eb264-db56-7835-27a2-634f0e6418d1, 'name': SearchDatastore_Task, 'duration_secs': 0.012735} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.842168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.842168] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.842168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.842168] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.842168] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.842843] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91eed578-3f52-463f-8a38-0d9828bd51ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.849248] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 815.849248] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5206a2cd-5cfb-b7bf-e0d6-cc35bcf41a0f" [ 815.849248] env[61964]: _type = "Task" [ 815.849248] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.858675] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5206a2cd-5cfb-b7bf-e0d6-cc35bcf41a0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.018088] env[61964]: DEBUG nova.compute.manager [None req-4964c045-0291-4265-8001-a39a9d0c969a tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Found 2 images (rotation: 2) {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4985}} [ 816.044570] env[61964]: DEBUG nova.network.neutron [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updated VIF entry in instance network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 816.044840] env[61964]: DEBUG nova.network.neutron [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.128973] env[61964]: DEBUG oslo_vmware.api [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041005, 'name': PowerOnVM_Task, 'duration_secs': 0.734547} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.129617] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.130741] env[61964]: INFO nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Took 9.89 seconds to spawn the instance on the hypervisor. [ 816.130741] env[61964]: DEBUG nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 816.131542] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d1bf3e-d9b0-4bd3-a187-c82dd36deb95 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.187112] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Releasing lock "refresh_cache-8bf586a2-0b56-4f5a-9f33-d537fba00d32" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.187621] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 816.187621] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing instance network info cache due to event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 816.187825] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.360109] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5206a2cd-5cfb-b7bf-e0d6-cc35bcf41a0f, 'name': SearchDatastore_Task, 'duration_secs': 0.00975} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.361224] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-551fccad-f512-4090-aeb8-2dd16828c6f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.371066] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 816.371066] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52656e2c-993c-b48a-6df7-12dfb146c7b9" [ 816.371066] env[61964]: _type = "Task" [ 816.371066] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.379252] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52656e2c-993c-b48a-6df7-12dfb146c7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.394264] env[61964]: DEBUG nova.network.neutron [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.422841] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.423087] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.431304] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f35bab-f148-491f-956f-b7c75079359b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.443624] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03134e4-8bb9-40a3-b83d-ba26ed2a179d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.481160] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b73675-5ce0-46ef-9fdb-6b1820f9f75c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.487228] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21d25bf-0631-4f4a-bd42-015bf9bbff0f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.503399] env[61964]: DEBUG nova.compute.provider_tree [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.519214] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.519462] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.550120] env[61964]: DEBUG oslo_concurrency.lockutils [req-f6962224-3a55-447d-899d-461c385e7219 req-3a2ae1c8-db34-46ff-9e36-42c5ef3c529a service nova] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.550563] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.550759] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.652289] env[61964]: INFO nova.compute.manager [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Took 45.21 seconds to build instance. [ 816.887020] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52656e2c-993c-b48a-6df7-12dfb146c7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.011861} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.887020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.887020] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 8bf586a2-0b56-4f5a-9f33-d537fba00d32/8bf586a2-0b56-4f5a-9f33-d537fba00d32.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 816.887020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.887020] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.887020] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a54402d8-b576-43f9-a826-63b667f6f4d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.891347] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35645f09-faba-4a59-a004-8ee79b19eff6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.899837] env[61964]: INFO nova.compute.manager [-] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Took 1.45 seconds to deallocate network for instance. [ 816.900776] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 816.900776] env[61964]: value = "task-1041006" [ 816.900776] env[61964]: _type = "Task" [ 816.900776] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.904650] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.904846] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 816.908549] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba5a1ccc-d917-40f9-9176-9e40309ed265 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.915701] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041006, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.917102] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 816.917102] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226ccdc-d4a7-01a6-c161-5b16d97b0600" [ 816.917102] env[61964]: _type = "Task" [ 816.917102] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.924572] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226ccdc-d4a7-01a6-c161-5b16d97b0600, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.926018] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 817.006454] env[61964]: DEBUG nova.scheduler.client.report [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 817.022130] env[61964]: DEBUG nova.compute.utils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.155275] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5a4b0322-a104-4cce-a14f-d3f3f1b5d4e6 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.718s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.410291] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.417945] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041006, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.431620] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226ccdc-d4a7-01a6-c161-5b16d97b0600, 'name': SearchDatastore_Task, 'duration_secs': 0.011498} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.436244] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0b87bd0-aed7-4620-ba52-a3e766927360 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.444357] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 817.444357] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528c477d-5af1-44ae-17e5-fac557a986e6" [ 817.444357] env[61964]: _type = "Task" [ 817.444357] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.455914] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528c477d-5af1-44ae-17e5-fac557a986e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.457133] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.511946] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.512600] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 817.515394] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.641s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.515627] env[61964]: DEBUG nova.objects.instance [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lazy-loading 'resources' on Instance uuid 58f8d368-fa0d-4c8d-9009-e6e1c5940032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.526483] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.793739] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updated VIF entry in instance network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.794148] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.920019] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041006, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644026} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.920019] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 8bf586a2-0b56-4f5a-9f33-d537fba00d32/8bf586a2-0b56-4f5a-9f33-d537fba00d32.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 817.920212] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.920403] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-122c62b6-edf4-411e-8d64-f30e89a87c7e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.929264] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 817.929264] env[61964]: value = "task-1041007" [ 817.929264] env[61964]: _type = "Task" [ 817.929264] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.943757] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041007, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.957150] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528c477d-5af1-44ae-17e5-fac557a986e6, 'name': SearchDatastore_Task, 'duration_secs': 0.043831} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.957571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.958015] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.958399] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-399009df-a437-4657-ac97-b8877894c4e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.966233] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 817.966233] env[61964]: value = "task-1041008" [ 817.966233] env[61964]: _type = "Task" [ 817.966233] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.982300] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.022077] env[61964]: DEBUG nova.compute.utils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.024046] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 818.027020] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 818.102069] env[61964]: DEBUG nova.policy [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c29d89b14fc4a12bf6b35e94b31373e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6110872ae3dc4491bb10abb1945ffe2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.297351] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.297963] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 818.298363] env[61964]: DEBUG nova.compute.manager [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing instance network info cache due to event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 818.298921] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquiring lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.299210] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Acquired lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.299416] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.396238] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff2276a-b45c-42cd-bb69-4c2a824c2e33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.401625] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca77bd03-a00e-453b-a566-7de40b14f61a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.446035] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ba229e-23a0-4737-8ef5-7fcc21002b78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.459565] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041007, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170493} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.464386] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.465453] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea887a7d-cf1b-4f37-a8ee-4cdceb7b1bd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.469822] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c64475e-edee-475f-85f1-22ff50011f53 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.482177] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.507772] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 8bf586a2-0b56-4f5a-9f33-d537fba00d32/8bf586a2-0b56-4f5a-9f33-d537fba00d32.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.508426] env[61964]: DEBUG nova.compute.provider_tree [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.509982] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e0f77bd-2fac-4aae-a7aa-850075e5cd5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.527923] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 818.535318] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 818.535318] env[61964]: value = "task-1041009" [ 818.535318] env[61964]: _type = "Task" [ 818.535318] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.542064] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041009, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.625701] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.626161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.626473] env[61964]: INFO nova.compute.manager [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Attaching volume a9253389-533d-4822-88cf-2d7417a75289 to /dev/sdb [ 818.675026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6585b3-f02d-420c-9ecc-427f0fd9ffe9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.683112] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac69e3e-cf30-4f8e-b5fe-419fff3aa10a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.699089] env[61964]: DEBUG nova.virt.block_device [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Updating existing volume attachment record: ac9388fb-8244-4601-b64d-074079d1c792 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 818.785749] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Successfully created port: 3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.912204] env[61964]: DEBUG nova.compute.manager [req-fa8341d4-7133-4f43-97d0-c060f07925fb req-6f412d13-e8ba-48c8-a0de-28eca81fd4c6 service nova] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Received event network-vif-deleted-6e60e0a5-9e9f-43f3-8823-cc1d7c0be746 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 818.986287] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.026453] env[61964]: DEBUG nova.scheduler.client.report [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.047211] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041009, 'name': ReconfigVM_Task, 'duration_secs': 0.385687} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.048231] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 8bf586a2-0b56-4f5a-9f33-d537fba00d32/8bf586a2-0b56-4f5a-9f33-d537fba00d32.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.048887] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-429da9f4-885f-4787-89d5-7a8f45a62db5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.055258] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 819.055258] env[61964]: value = "task-1041013" [ 819.055258] env[61964]: _type = "Task" [ 819.055258] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.063894] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041013, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.155109] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updated VIF entry in instance network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 819.155840] env[61964]: DEBUG nova.network.neutron [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [{"id": "1c91d247-4361-482e-a473-d6cc5926a4ca", "address": "fa:16:3e:27:99:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c91d247-43", "ovs_interfaceid": "1c91d247-4361-482e-a473-d6cc5926a4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.214425] env[61964]: DEBUG nova.compute.manager [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 819.214628] env[61964]: DEBUG nova.compute.manager [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing instance network info cache due to event network-changed-1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 819.214869] env[61964]: DEBUG oslo_concurrency.lockutils [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] Acquiring lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.485927] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041008, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.534329] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.019s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.538576] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.935s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.540890] env[61964]: INFO nova.compute.claims [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.547298] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 819.573909] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041013, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.577624] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 819.577913] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.578090] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 819.578585] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.578747] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 819.578911] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 819.579151] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 819.579379] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 819.579583] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 819.579759] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 819.579939] env[61964]: DEBUG nova.virt.hardware [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 819.581050] env[61964]: INFO nova.scheduler.client.report [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Deleted allocations for instance 58f8d368-fa0d-4c8d-9009-e6e1c5940032 [ 819.582792] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4504977-2595-4c2a-ae48-e86c85b90e4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.594344] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57863755-61df-42ba-96de-fd96c909f498 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.658763] env[61964]: DEBUG oslo_concurrency.lockutils [req-674e2a22-91ab-4c6c-9829-f276942b1009 req-f4a86d74-5406-4466-bea9-8b220491e81f service nova] Releasing lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.659920] env[61964]: DEBUG oslo_concurrency.lockutils [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] Acquired lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.659920] env[61964]: DEBUG nova.network.neutron [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Refreshing network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 819.781413] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.781762] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.782143] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.782143] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.782333] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.784631] env[61964]: INFO nova.compute.manager [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Terminating instance [ 819.985595] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041008, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.872594} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.985750] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 819.985961] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.986232] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc941480-44cb-4765-8ff3-eebe2967d061 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.992554] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 819.992554] env[61964]: value = "task-1041014" [ 819.992554] env[61964]: _type = "Task" [ 819.992554] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.000233] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.067891] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041013, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.092022] env[61964]: DEBUG oslo_concurrency.lockutils [None req-66defdfa-a091-43d4-ab09-28b68b793bf6 tempest-ServersTestJSON-1702022353 tempest-ServersTestJSON-1702022353-project-member] Lock "58f8d368-fa0d-4c8d-9009-e6e1c5940032" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.613s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.224018] env[61964]: DEBUG nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 820.224018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb1f830-0e0c-4937-a29d-32eb50ce925b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.230647] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.231053] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.290235] env[61964]: DEBUG nova.compute.manager [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 820.290650] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 820.291302] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659db109-7afa-4467-a0e1-e8f9f921fe6e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.300200] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 820.300653] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-576800b0-2445-4487-b025-d9e694233200 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.311016] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 820.311016] env[61964]: value = "task-1041015" [ 820.311016] env[61964]: _type = "Task" [ 820.311016] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.318187] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.507678] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.539160] env[61964]: DEBUG nova.network.neutron [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updated VIF entry in instance network info cache for port 1c91d247-4361-482e-a473-d6cc5926a4ca. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.539541] env[61964]: DEBUG nova.network.neutron [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [{"id": "1c91d247-4361-482e-a473-d6cc5926a4ca", "address": "fa:16:3e:27:99:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c91d247-43", "ovs_interfaceid": "1c91d247-4361-482e-a473-d6cc5926a4ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.556268] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "2a185e4b-c636-418a-a75e-2cf158f550c0" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.556268] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.556268] env[61964]: INFO nova.compute.manager [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Rebooting instance [ 820.568466] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041013, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.736637] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 820.741164] env[61964]: INFO nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] instance snapshotting [ 820.741164] env[61964]: DEBUG nova.objects.instance [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.821649] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041015, 'name': PowerOffVM_Task, 'duration_secs': 0.477846} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.821649] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 820.821826] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 820.821967] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc94bfbd-8140-40bb-8c06-7da3e4731649 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.864294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9826549d-058f-4839-b2d3-a4a0fa27fcc9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.871798] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0e5aed-c17c-4132-82fa-0c15d2c06f14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.902724] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94d7bec-efc2-4e7c-aa15-e5425f0944fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.905437] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 820.905637] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 820.905820] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleting the datastore file [datastore2] ab33fb75-4114-4a26-8692-befb407b7460 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.906083] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b524376c-d1d4-499d-998a-67418055a89a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.915762] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8392fe2-7e70-46f7-ae53-86cc809bbb22 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.919942] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 820.919942] env[61964]: value = "task-1041017" [ 820.919942] env[61964]: _type = "Task" [ 820.919942] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.930828] env[61964]: DEBUG nova.compute.provider_tree [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.936394] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.990567] env[61964]: DEBUG oslo_concurrency.lockutils [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.990736] env[61964]: DEBUG oslo_concurrency.lockutils [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.991062] env[61964]: DEBUG nova.objects.instance [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'flavor' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.007118] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.681782} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.007336] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.008196] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd89ccd6-c953-45d0-a85f-0f5039c999c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.030529] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.031599] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14e3d17c-5ee6-4488-ac0b-146cf106d63f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.046906] env[61964]: DEBUG oslo_concurrency.lockutils [req-e98bc1c3-b2c7-4c59-83fd-c310d85952fa req-8bde1fb8-07ea-4482-abcf-fdd367ee835f service nova] Releasing lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.048173] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Successfully updated port: 3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.054174] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 821.054174] env[61964]: value = "task-1041018" [ 821.054174] env[61964]: _type = "Task" [ 821.054174] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.067014] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041018, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.072453] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041013, 'name': Rename_Task, 'duration_secs': 1.676684} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.072715] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 821.072956] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7fc6a52-acb9-4f38-937e-cea5aa32dbec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.078998] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 821.078998] env[61964]: value = "task-1041019" [ 821.078998] env[61964]: _type = "Task" [ 821.078998] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.088047] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.146958] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.147375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquired lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.147638] env[61964]: DEBUG nova.network.neutron [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.249950] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf433a4-6968-4bdb-98ff-d69b0d06be98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.269563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.270431] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a4ce4c-f3e3-4252-803f-f8145f35ac36 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.434411] env[61964]: DEBUG nova.scheduler.client.report [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 821.439342] env[61964]: DEBUG oslo_vmware.api [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.386667} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.439745] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.440015] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 821.440941] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 821.440941] env[61964]: INFO nova.compute.manager [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 1.15 seconds to destroy the instance on the hypervisor. [ 821.440941] env[61964]: DEBUG oslo.service.loopingcall [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.441105] env[61964]: DEBUG nova.compute.manager [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 821.441229] env[61964]: DEBUG nova.network.neutron [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.495618] env[61964]: DEBUG nova.objects.instance [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'pci_requests' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.550686] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.550957] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.550996] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.552918] env[61964]: DEBUG nova.compute.manager [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Received event network-vif-plugged-3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 821.553186] env[61964]: DEBUG oslo_concurrency.lockutils [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.553412] env[61964]: DEBUG oslo_concurrency.lockutils [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.553595] env[61964]: DEBUG oslo_concurrency.lockutils [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.553766] env[61964]: DEBUG nova.compute.manager [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] No waiting events found dispatching network-vif-plugged-3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 821.553937] env[61964]: WARNING nova.compute.manager [req-4558cbff-95fb-4338-8008-4955ef2d3dde req-06c2bb4a-b199-4d21-ac00-5dd77d495ddc service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Received unexpected event network-vif-plugged-3a986bcd-6584-4440-bf01-7db0dff75b2f for instance with vm_state building and task_state spawning. [ 821.564305] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041018, 'name': ReconfigVM_Task, 'duration_secs': 0.489826} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.564611] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Reconfigured VM instance instance-0000003f to attach disk [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1/a23e7f40-1b3e-4c8f-b664-30f6314923a1.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.565257] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a69a8803-3c84-4bfd-9f0b-8a240c9a4953 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.572518] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 821.572518] env[61964]: value = "task-1041020" [ 821.572518] env[61964]: _type = "Task" [ 821.572518] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.581797] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041020, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.591330] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041019, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.783274] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 821.783274] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2af6e72d-bcd8-4496-9a91-83cf1134f606 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.790199] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 821.790199] env[61964]: value = "task-1041021" [ 821.790199] env[61964]: _type = "Task" [ 821.790199] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.803843] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041021, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.807914] env[61964]: DEBUG nova.compute.manager [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 821.808052] env[61964]: DEBUG nova.compute.manager [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing instance network info cache due to event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 821.808263] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Acquiring lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.947040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.947040] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 821.953157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.530s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.974935] env[61964]: DEBUG nova.network.neutron [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.998164] env[61964]: DEBUG nova.objects.base [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 821.998401] env[61964]: DEBUG nova.network.neutron [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.082658] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041020, 'name': Rename_Task, 'duration_secs': 0.178052} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.089130] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.090265] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b144b1e0-8e89-4e8a-a3a3-7fe254cdad9b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.102393] env[61964]: DEBUG oslo_vmware.api [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041019, 'name': PowerOnVM_Task, 'duration_secs': 0.744989} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.107588] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.109190] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.109465] env[61964]: INFO nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Took 11.20 seconds to spawn the instance on the hypervisor. [ 822.109601] env[61964]: DEBUG nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 822.110643] env[61964]: DEBUG nova.compute.manager [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Received event network-changed-3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 822.110864] env[61964]: DEBUG nova.compute.manager [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Refreshing instance network info cache due to event network-changed-3a986bcd-6584-4440-bf01-7db0dff75b2f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 822.111232] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] Acquiring lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.111770] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 822.111770] env[61964]: value = "task-1041022" [ 822.111770] env[61964]: _type = "Task" [ 822.111770] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.113123] env[61964]: DEBUG oslo_concurrency.lockutils [None req-15f894da-721c-4efd-bf49-bbafa66756c8 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.122s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.114814] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bd2fb0-c14f-4dad-9f33-54a71f4517d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.135236] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041022, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.196596] env[61964]: DEBUG nova.network.neutron [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.300484] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041021, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.317797] env[61964]: DEBUG nova.network.neutron [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.348105] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37effc2c-54e6-404f-8147-07022628ec37 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.358075] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0abced9-3cf9-4b84-ab9d-5c444a30773b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.389523] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004ca1fb-b419-4c06-95f2-9844efc82eab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.398541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4fe176-48cb-4cf3-9121-dfe883bd7c20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.413652] env[61964]: DEBUG nova.compute.provider_tree [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.460235] env[61964]: DEBUG nova.compute.utils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.461603] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 822.461777] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.477842] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Releasing lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.479161] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Acquired lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.479439] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.502737] env[61964]: DEBUG nova.policy [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4bf81d972234720b707312887c1b74f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f864c21d43647f194c05c358e2dc66c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.630523] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041022, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.642039] env[61964]: INFO nova.compute.manager [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Took 31.71 seconds to build instance. [ 822.698842] env[61964]: INFO nova.compute.manager [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 1.26 seconds to deallocate network for instance. [ 822.758143] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Successfully created port: 19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.802016] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041021, 'name': CreateSnapshot_Task, 'duration_secs': 0.546162} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.802785] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 822.803632] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e971c745-e5e7-4cdb-9c4a-03682ad8b6f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.818018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.818018] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Instance network_info: |[{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 822.818018] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] Acquired lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.818018] env[61964]: DEBUG nova.network.neutron [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Refreshing network info cache for port 3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.818839] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:51:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a986bcd-6584-4440-bf01-7db0dff75b2f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.827100] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Creating folder: Project (6110872ae3dc4491bb10abb1945ffe2a). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.831019] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61209208-fd4c-4d1c-84c9-c8e4785da2e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.842264] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Created folder: Project (6110872ae3dc4491bb10abb1945ffe2a) in parent group-v230360. [ 822.842464] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Creating folder: Instances. Parent ref: group-v230483. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.842715] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c53e864c-d4d7-4287-b7d8-849264e511d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.852793] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Created folder: Instances in parent group-v230483. [ 822.853067] env[61964]: DEBUG oslo.service.loopingcall [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.853511] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.853757] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65c7793e-5316-4ec0-9fd4-392a2c8eedbd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.874526] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.874526] env[61964]: value = "task-1041025" [ 822.874526] env[61964]: _type = "Task" [ 822.874526] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.882822] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041025, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.916442] env[61964]: DEBUG nova.scheduler.client.report [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 822.965322] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 822.985534] env[61964]: DEBUG nova.compute.manager [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 822.987014] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846d53b3-dc21-477d-bbd7-a45d22358a4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.135223] env[61964]: DEBUG oslo_vmware.api [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041022, 'name': PowerOnVM_Task, 'duration_secs': 1.007451} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.135428] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.135655] env[61964]: DEBUG nova.compute.manager [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 823.136456] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691cabd6-ccc1-4098-8ff1-e0dbb239052f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.147146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cb0d0ee9-3108-46f7-876d-bd490473c03c tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.218s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.209951] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.323544] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 823.332596] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6b2fac4f-335b-4768-8b3f-c00d3c4b366b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.347125] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 823.347125] env[61964]: value = "task-1041027" [ 823.347125] env[61964]: _type = "Task" [ 823.347125] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.357413] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041027, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.390055] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041025, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.425506] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.469s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.425506] env[61964]: INFO nova.compute.manager [None req-eacf22f6-5769-4df0-a32b-df9bfb04d071 tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Successfully reverted task state from rebuilding on failure for instance. [ 823.431714] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.997s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.434443] env[61964]: INFO nova.compute.claims [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.665328] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.717419] env[61964]: DEBUG nova.network.neutron [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updated VIF entry in instance network info cache for port 3a986bcd-6584-4440-bf01-7db0dff75b2f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.717907] env[61964]: DEBUG nova.network.neutron [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.769799] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updated VIF entry in instance network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.770201] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.859020] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041027, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.887860] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041025, 'name': CreateVM_Task, 'duration_secs': 0.598877} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.888117] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 823.888820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.889039] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.889376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.889711] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6936375d-83fb-474f-87ac-0581b0020769 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.894955] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 823.894955] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528a41ef-f25d-d5a2-86bd-11870d58f1b1" [ 823.894955] env[61964]: _type = "Task" [ 823.894955] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.905386] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528a41ef-f25d-d5a2-86bd-11870d58f1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.978049] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 824.007100] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684bfc6b-5161-42a5-9711-63f5810f5560 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.016950] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 824.017319] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.017588] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 824.017907] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.018204] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 824.018460] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 824.018777] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 824.018979] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 824.020299] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 824.023019] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 824.023019] env[61964]: DEBUG nova.virt.hardware [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 824.023019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace27670-40e1-400b-9605-b6f3687035b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.033315] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Doing hard reboot of VM {{(pid=61964) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 824.037535] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-6bc7078c-430d-434e-9141-699734b69bad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.041028] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ad391e-b574-461f-911f-c932e898cab6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.061656] env[61964]: DEBUG oslo_vmware.api [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 824.061656] env[61964]: value = "task-1041028" [ 824.061656] env[61964]: _type = "Task" [ 824.061656] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.071708] env[61964]: DEBUG oslo_vmware.api [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041028, 'name': ResetVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.221473] env[61964]: DEBUG oslo_concurrency.lockutils [req-4d53c579-3d41-4e9a-9f07-c681119d8fa4 req-cc95cf88-1b33-4210-bb6a-b772ebfe04b5 service nova] Releasing lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.275520] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Releasing lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.275520] env[61964]: DEBUG nova.compute.manager [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 824.275520] env[61964]: DEBUG nova.compute.manager [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing instance network info cache due to event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 824.275520] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.275520] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.275520] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.360026] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041027, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.408019] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528a41ef-f25d-d5a2-86bd-11870d58f1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.011668} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.408481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.408795] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.409449] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.409715] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.409973] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.410278] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e551f3e0-8932-4568-b0d7-3737920dd980 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.421107] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.421642] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.422703] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ea1b53c-b429-4f19-9224-5948d5702f61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.432257] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 824.432257] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ec124-9629-d5ce-eb22-52d064fee26c" [ 824.432257] env[61964]: _type = "Task" [ 824.432257] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.442214] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ec124-9629-d5ce-eb22-52d064fee26c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.584494] env[61964]: DEBUG oslo_vmware.api [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041028, 'name': ResetVM_Task, 'duration_secs': 0.09985} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.584494] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Did hard reboot of VM {{(pid=61964) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 824.584494] env[61964]: DEBUG nova.compute.manager [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 824.584494] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa8a1a4-be2f-49fb-9aba-0e7f6fcfab9e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.732593] env[61964]: DEBUG nova.compute.manager [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 824.733291] env[61964]: DEBUG nova.compute.manager [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing instance network info cache due to event network-changed-89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 824.733586] env[61964]: DEBUG oslo_concurrency.lockutils [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] Acquiring lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.734979] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Successfully updated port: 19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.858980] env[61964]: DEBUG nova.compute.manager [req-d1f9811c-c9ea-4e42-806a-121cccd198d6 req-bf169621-a2e9-45c9-811d-f2359c3cf61b service nova] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Received event network-vif-deleted-1c91d247-4361-482e-a473-d6cc5926a4ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 824.863054] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041027, 'name': CloneVM_Task, 'duration_secs': 1.402063} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.864183] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Created linked-clone VM from snapshot [ 824.864957] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf074497-7967-4653-b202-77ba727c82ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.871849] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501b6473-4f83-4c84-a1ce-d95d3fd85078 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.883028] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploading image f04b1f74-105c-4cda-9f12-0aea28fe5e43 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 824.885735] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada6c600-e10c-4725-b0ce-a7366f950f50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.924913] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8d8c59-4986-4d22-8e99-5cb023782ce4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.935229] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 824.935229] env[61964]: value = "vm-230486" [ 824.935229] env[61964]: _type = "VirtualMachine" [ 824.935229] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 824.937354] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-08ee3e11-6cb8-4267-8e5c-3bf05d47da05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.942703] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4b0695-8425-4753-bcd2-2d8139eaf03c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.950968] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ec124-9629-d5ce-eb22-52d064fee26c, 'name': SearchDatastore_Task, 'duration_secs': 0.016448} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.952727] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02483e4a-cebd-481d-b5a3-2bf442c7550f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.963804] env[61964]: DEBUG nova.compute.provider_tree [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.965508] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease: (returnval){ [ 824.965508] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5271052b-4229-4cad-f972-9ce5aef99423" [ 824.965508] env[61964]: _type = "HttpNfcLease" [ 824.965508] env[61964]: } obtained for exporting VM: (result){ [ 824.965508] env[61964]: value = "vm-230486" [ 824.965508] env[61964]: _type = "VirtualMachine" [ 824.965508] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 824.965733] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the lease: (returnval){ [ 824.965733] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5271052b-4229-4cad-f972-9ce5aef99423" [ 824.965733] env[61964]: _type = "HttpNfcLease" [ 824.965733] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 824.970191] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 824.970191] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527bbb2a-08dd-8dac-3104-3c2631d40da2" [ 824.970191] env[61964]: _type = "Task" [ 824.970191] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.978129] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 824.978129] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5271052b-4229-4cad-f972-9ce5aef99423" [ 824.978129] env[61964]: _type = "HttpNfcLease" [ 824.978129] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 824.978637] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 824.978637] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5271052b-4229-4cad-f972-9ce5aef99423" [ 824.978637] env[61964]: _type = "HttpNfcLease" [ 824.978637] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 824.979399] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351cd30f-bdf3-414c-a5b1-9a69c5922bff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.984751] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527bbb2a-08dd-8dac-3104-3c2631d40da2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.990195] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 824.990325] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 825.086072] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8217e693-8b00-4429-9a02-0d78d9b94a57 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.103500] env[61964]: DEBUG oslo_concurrency.lockutils [None req-94a1694e-66d1-4f55-83ad-0c41bc374362 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.547s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.183856] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.184171] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.184559] env[61964]: DEBUG nova.objects.instance [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'flavor' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.237546] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.237735] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.241020] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 825.244721] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updated VIF entry in instance network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 825.245048] env[61964]: DEBUG nova.network.neutron [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.272113] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 825.272164] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230481', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'name': 'volume-a9253389-533d-4822-88cf-2d7417a75289', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c922228-4ba9-43ed-8e5c-3ca15e48ceab', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'serial': 'a9253389-533d-4822-88cf-2d7417a75289'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 825.273089] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a003b42-d65c-4e00-88f3-09de9de3f0c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.294582] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8e6511-d120-496b-9a03-fb6d6008908e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.325598] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] volume-a9253389-533d-4822-88cf-2d7417a75289/volume-a9253389-533d-4822-88cf-2d7417a75289.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.326544] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd972243-8def-4a9d-952b-69a23c8ef795 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.340155] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.346677] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 825.346677] env[61964]: value = "task-1041030" [ 825.346677] env[61964]: _type = "Task" [ 825.346677] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.355840] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.467732] env[61964]: DEBUG nova.scheduler.client.report [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.484976] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527bbb2a-08dd-8dac-3104-3c2631d40da2, 'name': SearchDatastore_Task, 'duration_secs': 0.026101} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.485940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.486266] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 138f44e4-e12e-4f89-a9b2-8a512b53cdf5/138f44e4-e12e-4f89-a9b2-8a512b53cdf5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 825.486564] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b3b3dc7-71b1-4aa6-b5ba-9cec6a3537bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.494642] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 825.494642] env[61964]: value = "task-1041031" [ 825.494642] env[61964]: _type = "Task" [ 825.494642] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.503867] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.752430] env[61964]: DEBUG oslo_concurrency.lockutils [req-774005b6-afbe-4ec6-91b7-6ba30f6f0cf4 req-40e7f64c-70c1-49c6-ad0a-8d578ece7652 service nova] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.759694] env[61964]: DEBUG oslo_concurrency.lockutils [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] Acquired lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.760510] env[61964]: DEBUG nova.network.neutron [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Refreshing network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.812790] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.829239] env[61964]: DEBUG nova.objects.instance [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'pci_requests' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.861500] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041030, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.975877] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.975877] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 825.978454] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.127s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.980190] env[61964]: INFO nova.compute.claims [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.012899] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041031, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.014652] env[61964]: DEBUG nova.network.neutron [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Updating instance_info_cache with network_info: [{"id": "19290ca0-b730-490e-b3a8-5a598373bcc7", "address": "fa:16:3e:d3:9c:ba", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19290ca0-b7", "ovs_interfaceid": "19290ca0-b730-490e-b3a8-5a598373bcc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.332657] env[61964]: DEBUG nova.objects.base [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 826.332887] env[61964]: DEBUG nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.359969] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041030, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.427341] env[61964]: DEBUG nova.policy [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.489867] env[61964]: DEBUG nova.compute.utils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.491682] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 826.518855] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612624} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.519633] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.520366] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Instance network_info: |[{"id": "19290ca0-b730-490e-b3a8-5a598373bcc7", "address": "fa:16:3e:d3:9c:ba", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19290ca0-b7", "ovs_interfaceid": "19290ca0-b730-490e-b3a8-5a598373bcc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 826.521032] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 138f44e4-e12e-4f89-a9b2-8a512b53cdf5/138f44e4-e12e-4f89-a9b2-8a512b53cdf5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.521926] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.523244] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:9c:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bab6a6c3-1c5c-4776-b21b-dec21196d702', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19290ca0-b730-490e-b3a8-5a598373bcc7', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.538931] env[61964]: DEBUG oslo.service.loopingcall [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.539444] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c41440f4-0471-4b4c-a356-3a65bb7bc407 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.542437] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 826.542757] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e24155b2-ac98-41da-a852-692d36383394 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.558944] env[61964]: DEBUG nova.network.neutron [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updated VIF entry in instance network info cache for port 89204ea4-fd9b-4979-8aaf-f224046531fa. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.559405] env[61964]: DEBUG nova.network.neutron [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [{"id": "89204ea4-fd9b-4979-8aaf-f224046531fa", "address": "fa:16:3e:1f:d3:b0", "network": {"id": "031529f1-1754-4934-be54-e365c1e17edc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2091302248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "540083b7cc4a4b8aa25714afc2add861", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89204ea4-fd", "ovs_interfaceid": "89204ea4-fd9b-4979-8aaf-f224046531fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.565372] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 826.565372] env[61964]: value = "task-1041032" [ 826.565372] env[61964]: _type = "Task" [ 826.565372] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.572185] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.572185] env[61964]: value = "task-1041033" [ 826.572185] env[61964]: _type = "Task" [ 826.572185] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.580988] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041032, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.587103] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041033, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.774835] env[61964]: DEBUG nova.compute.manager [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 826.774835] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d61ae2-b337-441a-b0c3-43628fa1decb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.863419] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041030, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.979836] env[61964]: DEBUG nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Successfully created port: 15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.996124] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 827.062497] env[61964]: DEBUG oslo_concurrency.lockutils [req-6191e4d6-9c71-43ed-a80d-e98a43a57b31 req-ae427b53-988f-4ba5-8462-7a75d0dcf8d2 service nova] Releasing lock "refresh_cache-46435c68-f85a-4360-b2b5-6296afc33c3e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.080800] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041032, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143637} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.086257] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.088066] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4ceff0-e098-4e9f-81b8-acaebbcb1699 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.889918] env[61964]: INFO nova.compute.manager [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] instance snapshotting [ 827.893612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.893612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.893612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.893612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.893612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.906817] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 138f44e4-e12e-4f89-a9b2-8a512b53cdf5/138f44e4-e12e-4f89-a9b2-8a512b53cdf5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.910958] env[61964]: INFO nova.compute.manager [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Terminating instance [ 827.916093] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fefc6c16-ec31-4b33-8a29-082816999a2e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.930489] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041033, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.930780] env[61964]: WARNING oslo_vmware.common.loopingcall [-] task run outlasted interval by 0.350391 sec [ 827.931588] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ca56ab-3281-4a4e-a6e8-a63b0be6d2ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.945767] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041030, 'name': ReconfigVM_Task, 'duration_secs': 1.543871} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.962226] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfigured VM instance instance-00000041 to attach disk [datastore1] volume-a9253389-533d-4822-88cf-2d7417a75289/volume-a9253389-533d-4822-88cf-2d7417a75289.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.967606] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 827.967606] env[61964]: value = "task-1041034" [ 827.967606] env[61964]: _type = "Task" [ 827.967606] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.974953] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-485733ab-6481-4d5e-a653-76f7234d11b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.987766] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38a2b59-0f78-48ac-b1b9-4264d3497152 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.991318] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041033, 'name': CreateVM_Task, 'duration_secs': 0.613259} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.995972] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 827.997368] env[61964]: DEBUG nova.compute.manager [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Received event network-vif-plugged-19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 827.997659] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Acquiring lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.997901] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.998152] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.998405] env[61964]: DEBUG nova.compute.manager [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] No waiting events found dispatching network-vif-plugged-19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 827.998680] env[61964]: WARNING nova.compute.manager [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Received unexpected event network-vif-plugged-19290ca0-b730-490e-b3a8-5a598373bcc7 for instance with vm_state building and task_state spawning. [ 827.998856] env[61964]: DEBUG nova.compute.manager [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Received event network-changed-19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 827.999110] env[61964]: DEBUG nova.compute.manager [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Refreshing instance network info cache due to event network-changed-19290ca0-b730-490e-b3a8-5a598373bcc7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 827.999377] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Acquiring lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.999527] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Acquired lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.999744] env[61964]: DEBUG nova.network.neutron [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Refreshing network info cache for port 19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 828.002945] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.003151] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.003502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.004270] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f19ba6c-dcf0-42c1-860d-80ea0213e0b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.014682] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.021202] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 828.021202] env[61964]: value = "task-1041035" [ 828.021202] env[61964]: _type = "Task" [ 828.021202] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.030066] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 828.030066] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52741aa1-df55-5f2f-bab8-fe97778ee64f" [ 828.030066] env[61964]: _type = "Task" [ 828.030066] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.036805] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041035, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.045996] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52741aa1-df55-5f2f-bab8-fe97778ee64f, 'name': SearchDatastore_Task, 'duration_secs': 0.014202} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.049328] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.049328] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.049328] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.049328] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.049328] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.049328] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2ebd6ac-4330-4cac-8eef-38cecc403613 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.059960] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.060229] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 828.061226] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9702ffa-32a0-49c3-9550-0230f3c213da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.069643] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 828.069643] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c5a8b-00f4-0896-b3fa-bb8d70f6af70" [ 828.069643] env[61964]: _type = "Task" [ 828.069643] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.080447] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c5a8b-00f4-0896-b3fa-bb8d70f6af70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.298595] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5042e4-61c9-4d70-8089-d1f776e5bd1f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.311041] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93877e3-d134-413f-b979-efdddd40b365 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.346143] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ee845c-59bd-43d4-adf3-c37b757a57c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.355783] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60815f94-4e89-4ab4-b20b-b30dc12db8c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.371334] env[61964]: DEBUG nova.compute.provider_tree [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.414390] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 828.439642] env[61964]: DEBUG nova.compute.manager [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 828.440040] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.440962] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3129058c-bc87-4e23-a0ef-61285faaf255 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.446976] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 828.447274] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.447439] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 828.447623] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.447773] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 828.447946] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 828.448145] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 828.448355] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 828.448536] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 828.448726] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 828.448919] env[61964]: DEBUG nova.virt.hardware [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 828.449818] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807932a4-a3a3-4110-ba53-5e3ecc7960ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.459563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.463526] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e927e05-a93a-4df5-a23f-e803fc78517c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.466552] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c385f4de-b66a-4b05-b11c-91967e6396f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.484787] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.491097] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Creating folder: Project (03b1d87b23e848f9951db4bcf4250806). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.492830] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62320a0f-c4c7-4655-8685-ffd701e51d97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.497038] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 828.497038] env[61964]: value = "task-1041036" [ 828.497038] env[61964]: _type = "Task" [ 828.497038] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.515935] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041034, 'name': ReconfigVM_Task, 'duration_secs': 0.531503} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.515935] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.517522] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 138f44e4-e12e-4f89-a9b2-8a512b53cdf5/138f44e4-e12e-4f89-a9b2-8a512b53cdf5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.518387] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Created folder: Project (03b1d87b23e848f9951db4bcf4250806) in parent group-v230360. [ 828.518582] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Creating folder: Instances. Parent ref: group-v230488. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.519507] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a824562e-79e7-45e7-90cd-6c6df8ae2f60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.523289] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a11706b4-78e6-4125-9af7-ae92c2b72e3e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.527020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 828.527020] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-355ca1c7-d174-467f-924e-845ec665d37e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.535743] env[61964]: DEBUG oslo_vmware.api [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041035, 'name': ReconfigVM_Task, 'duration_secs': 0.218054} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.537303] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230481', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'name': 'volume-a9253389-533d-4822-88cf-2d7417a75289', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c922228-4ba9-43ed-8e5c-3ca15e48ceab', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'serial': 'a9253389-533d-4822-88cf-2d7417a75289'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 828.538961] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 828.538961] env[61964]: value = "task-1041038" [ 828.538961] env[61964]: _type = "Task" [ 828.538961] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.540970] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 828.540970] env[61964]: value = "task-1041039" [ 828.540970] env[61964]: _type = "Task" [ 828.540970] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.550303] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Created folder: Instances in parent group-v230488. [ 828.550303] env[61964]: DEBUG oslo.service.loopingcall [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.551032] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.551804] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa0c4108-abf2-4eb7-af50-a8424163b1a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.575066] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041038, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.575422] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041039, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.583314] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.583314] env[61964]: value = "task-1041041" [ 828.583314] env[61964]: _type = "Task" [ 828.583314] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.595412] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529c5a8b-00f4-0896-b3fa-bb8d70f6af70, 'name': SearchDatastore_Task, 'duration_secs': 0.012161} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.596747] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d38e495a-f4e2-40d9-b816-4536e444d423 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.603708] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041041, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.608029] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 828.608029] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52afa19c-b239-97cc-d562-06cf42217fe8" [ 828.608029] env[61964]: _type = "Task" [ 828.608029] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.616144] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52afa19c-b239-97cc-d562-06cf42217fe8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.800814] env[61964]: DEBUG nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Successfully updated port: 15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.833421] env[61964]: DEBUG nova.network.neutron [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Updated VIF entry in instance network info cache for port 19290ca0-b730-490e-b3a8-5a598373bcc7. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.833851] env[61964]: DEBUG nova.network.neutron [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Updating instance_info_cache with network_info: [{"id": "19290ca0-b730-490e-b3a8-5a598373bcc7", "address": "fa:16:3e:d3:9c:ba", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19290ca0-b7", "ovs_interfaceid": "19290ca0-b730-490e-b3a8-5a598373bcc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.874543] env[61964]: DEBUG nova.scheduler.client.report [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 829.009575] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041036, 'name': PowerOffVM_Task, 'duration_secs': 0.241333} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.009840] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.009980] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.010230] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-442e1e8f-87d3-483e-a397-73e58eb8aba1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.057019] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041038, 'name': Rename_Task, 'duration_secs': 0.198675} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.060307] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 829.060613] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041039, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.060951] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec9e5745-68ce-4804-b681-771424cf03dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.069810] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 829.069810] env[61964]: value = "task-1041043" [ 829.069810] env[61964]: _type = "Task" [ 829.069810] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.082192] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.099980] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041041, 'name': CreateVM_Task, 'duration_secs': 0.388771} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.099980] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.099980] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.099980] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.099980] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.099980] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4fa2f94-7253-4556-ab51-19dfd5f2ce22 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.114897] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.115920] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.115920] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] a23e7f40-1b3e-4c8f-b664-30f6314923a1 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.115920] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 829.115920] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c4efcb-e039-1d1e-5f87-4d68d440a8ed" [ 829.115920] env[61964]: _type = "Task" [ 829.115920] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.116996] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae0ea77f-a4b1-4f6c-bde0-98ad0ee14303 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.127604] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52afa19c-b239-97cc-d562-06cf42217fe8, 'name': SearchDatastore_Task, 'duration_secs': 0.01223} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.128783] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.128783] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] b90e7ebd-8dc6-495b-99bf-15ff24763af4/b90e7ebd-8dc6-495b-99bf-15ff24763af4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 829.129642] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42dcdb7f-1b65-4c06-ad6c-3a4b5cb61fb5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.137956] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c4efcb-e039-1d1e-5f87-4d68d440a8ed, 'name': SearchDatastore_Task, 'duration_secs': 0.016498} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.137956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.137956] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.137956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.137956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.138453] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.138453] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 829.138453] env[61964]: value = "task-1041044" [ 829.138453] env[61964]: _type = "Task" [ 829.138453] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.140222] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02b99508-a705-4003-a925-fcd5c772c8c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.142447] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 829.142447] env[61964]: value = "task-1041045" [ 829.142447] env[61964]: _type = "Task" [ 829.142447] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.155148] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.158439] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041044, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.164976] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.165618] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 829.166097] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99c4f43e-19e0-46b2-8188-c2d55e1b8d76 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.174784] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 829.174784] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a0c29a-91a1-5b92-5e9d-034135e75a99" [ 829.174784] env[61964]: _type = "Task" [ 829.174784] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.189478] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a0c29a-91a1-5b92-5e9d-034135e75a99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.303421] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.303630] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.303920] env[61964]: DEBUG nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.337318] env[61964]: DEBUG oslo_concurrency.lockutils [req-3d10b99f-9f00-453e-aeb9-c6e45f978c8e req-fa8e58c0-713d-4742-ac7c-c154b98f4e05 service nova] Releasing lock "refresh_cache-b90e7ebd-8dc6-495b-99bf-15ff24763af4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.337610] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.337996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.381746] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.403s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.382057] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 829.386041] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.061s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.386518] env[61964]: DEBUG nova.objects.instance [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lazy-loading 'resources' on Instance uuid 062da016-f4d6-424d-90e6-5903940e2daa {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.563816] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041039, 'name': CreateSnapshot_Task, 'duration_secs': 0.769639} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.565557] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 829.565557] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab298da-aa46-4764-9d2c-e69818c0dffe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.588542] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041043, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.594828] env[61964]: DEBUG nova.objects.instance [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'flavor' on Instance uuid 1c922228-4ba9-43ed-8e5c-3ca15e48ceab {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.658150] env[61964]: DEBUG oslo_vmware.api [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041044, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214692} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.662202] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.662583] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.662831] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.663073] env[61964]: INFO nova.compute.manager [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Took 1.22 seconds to destroy the instance on the hypervisor. [ 829.663347] env[61964]: DEBUG oslo.service.loopingcall [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.664586] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041045, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.664586] env[61964]: DEBUG nova.compute.manager [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 829.664737] env[61964]: DEBUG nova.network.neutron [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.695023] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a0c29a-91a1-5b92-5e9d-034135e75a99, 'name': SearchDatastore_Task, 'duration_secs': 0.011562} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.695023] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85e19500-5ee8-4be4-a886-c08c75b9d0ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.700529] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 829.700529] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aa0d31-bace-ccd7-0365-9f08c316c0be" [ 829.700529] env[61964]: _type = "Task" [ 829.700529] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.710153] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aa0d31-bace-ccd7-0365-9f08c316c0be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.840597] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 829.850031] env[61964]: WARNING nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] 48dd64e7-3659-4104-8273-6f58f8d3eb44 already exists in list: networks containing: ['48dd64e7-3659-4104-8273-6f58f8d3eb44']. ignoring it [ 829.888984] env[61964]: DEBUG nova.compute.utils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.890864] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 830.096011] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 830.103191] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9bd3f500-2e61-4bc6-bef7-b708fb6365df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.109091] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29353db4-bc47-4148-86a0-9a1afd057141 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 11.483s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.110643] env[61964]: DEBUG oslo_vmware.api [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041043, 'name': PowerOnVM_Task, 'duration_secs': 0.642643} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.111174] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.111408] env[61964]: INFO nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Took 10.56 seconds to spawn the instance on the hypervisor. [ 830.111595] env[61964]: DEBUG nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.114510] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d480f1-9ff8-4a1c-9eca-f595c1055fe4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.119990] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 830.119990] env[61964]: value = "task-1041046" [ 830.119990] env[61964]: _type = "Task" [ 830.119990] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.148083] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041046, 'name': CloneVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.158279] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602395} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.161258] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] b90e7ebd-8dc6-495b-99bf-15ff24763af4/b90e7ebd-8dc6-495b-99bf-15ff24763af4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 830.161650] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.162313] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6672f5aa-9858-4cfe-bb36-7a06293f14a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.173152] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 830.173152] env[61964]: value = "task-1041047" [ 830.173152] env[61964]: _type = "Task" [ 830.173152] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.187133] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041047, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.215126] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52aa0d31-bace-ccd7-0365-9f08c316c0be, 'name': SearchDatastore_Task, 'duration_secs': 0.027939} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.218101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.218504] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.219567] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef505a4b-b904-4d09-a5a7-fd7a9027b168 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.231306] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 830.231306] env[61964]: value = "task-1041048" [ 830.231306] env[61964]: _type = "Task" [ 830.231306] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.244093] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041048, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.350323] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c807eb-9cee-4e51-868f-a8c114b4c8b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.368901] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904c5ce2-fa38-44a5-94d4-42a5bae43897 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.374465] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.402248] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 830.407132] env[61964]: DEBUG nova.network.neutron [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15153388-0bbd-4e56-b179-42a3c5ad9552", "address": "fa:16:3e:86:91:41", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15153388-0b", "ovs_interfaceid": "15153388-0bbd-4e56-b179-42a3c5ad9552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.409201] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99071f08-0749-44d8-ae51-9d343d36f6f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.418719] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d45ee2b-3ea9-4009-a45d-1a6851efe88a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.437160] env[61964]: DEBUG nova.network.neutron [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.438758] env[61964]: DEBUG nova.compute.provider_tree [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.638112] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041046, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.641086] env[61964]: INFO nova.compute.manager [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Took 32.96 seconds to build instance. [ 830.688024] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041047, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.16684} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.688024] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.688024] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92de4d7e-5051-4be7-945d-b365cbee2d32 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.717097] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] b90e7ebd-8dc6-495b-99bf-15ff24763af4/b90e7ebd-8dc6-495b-99bf-15ff24763af4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.717572] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b184e7d-2932-4262-8131-262a4bb45f86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.747739] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041048, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.749685] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 830.749685] env[61964]: value = "task-1041049" [ 830.749685] env[61964]: _type = "Task" [ 830.749685] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.761583] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041049, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.914350] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.915614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.915938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.917759] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9c714c-f4b2-4c66-b8c2-f943866cb2c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.942035] env[61964]: INFO nova.compute.manager [-] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Took 1.28 seconds to deallocate network for instance. [ 830.942963] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 830.943460] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.943740] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 830.944091] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.944398] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 830.944719] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 830.945110] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 830.945409] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 830.945831] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 830.946171] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 830.946497] env[61964]: DEBUG nova.virt.hardware [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 830.955175] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfiguring VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 830.957714] env[61964]: DEBUG nova.scheduler.client.report [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 830.965232] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35fa4be0-f9df-4187-8970-7965862ab8b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.984505] env[61964]: DEBUG nova.compute.manager [req-d70f1f05-2577-422e-a535-f75d9a86c9f3 req-3d1f76c0-fa92-4e31-b7ca-79901189db6f service nova] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Received event network-vif-deleted-24e07767-885b-4cd5-9127-da8635ce0092 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 830.994586] env[61964]: DEBUG oslo_vmware.api [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 830.994586] env[61964]: value = "task-1041050" [ 830.994586] env[61964]: _type = "Task" [ 830.994586] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.006772] env[61964]: DEBUG oslo_vmware.api [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041050, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.088745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "2a185e4b-c636-418a-a75e-2cf158f550c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.089212] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.089504] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.089693] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.089994] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.094055] env[61964]: INFO nova.compute.manager [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Terminating instance [ 831.136512] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041046, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.149241] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a72248a9-9234-4f3e-9f32-02b32c655c63 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.483s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.149681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "a1e5c369-4233-4794-adda-1ba1e89e8154" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.150098] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.161154] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 831.161154] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing instance network info cache due to event network-changed-df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 831.161154] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Acquiring lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.161154] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Acquired lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.161154] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Refreshing network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.250023] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041048, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663912} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.250633] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.251041] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.256399] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86006db5-9c94-48ba-b559-6606a66737d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.269654] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041049, 'name': ReconfigVM_Task, 'duration_secs': 0.517884} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.270634] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Reconfigured VM instance instance-00000045 to attach disk [datastore1] b90e7ebd-8dc6-495b-99bf-15ff24763af4/b90e7ebd-8dc6-495b-99bf-15ff24763af4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.271711] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 831.271711] env[61964]: value = "task-1041051" [ 831.271711] env[61964]: _type = "Task" [ 831.271711] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.272390] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-378542a1-f179-465c-afaa-a4cf50c514fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.288157] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.290511] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 831.290511] env[61964]: value = "task-1041052" [ 831.290511] env[61964]: _type = "Task" [ 831.290511] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.305636] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041052, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.370771] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.371183] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.371491] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.371707] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.371888] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.374278] env[61964]: INFO nova.compute.manager [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Terminating instance [ 831.418253] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 831.453211] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 831.453876] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.453876] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 831.454583] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.454583] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 831.454583] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 831.454583] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 831.454841] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 831.455723] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 831.455723] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 831.455723] env[61964]: DEBUG nova.virt.hardware [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 831.457231] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d9cc82-5013-43c8-a6c5-a46cc391af7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.467429] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b644b7b-5843-452d-a459-b76a25853827 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.485720] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.100s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.488157] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.494135] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Creating folder: Project (edd34978a2064fee91a2f73ba6e0b2cc). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.494785] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.983s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.495076] env[61964]: DEBUG nova.objects.instance [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lazy-loading 'resources' on Instance uuid 560bd013-eae1-4541-aaa2-0732d8f20e4a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.496983] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.497257] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-956bdcef-a678-47f4-b7c2-6c712f879d6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.510847] env[61964]: DEBUG oslo_vmware.api [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041050, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.517747] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Created folder: Project (edd34978a2064fee91a2f73ba6e0b2cc) in parent group-v230360. [ 831.517747] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Creating folder: Instances. Parent ref: group-v230493. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 831.517982] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a38b1086-0e56-4dd7-9882-a0f4e25a9e26 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.523724] env[61964]: INFO nova.scheduler.client.report [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted allocations for instance 062da016-f4d6-424d-90e6-5903940e2daa [ 831.531486] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Created folder: Instances in parent group-v230493. [ 831.531886] env[61964]: DEBUG oslo.service.loopingcall [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.532197] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.532488] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca09ca30-e32c-49ca-8ed8-4459f631139c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.554343] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.554343] env[61964]: value = "task-1041055" [ 831.554343] env[61964]: _type = "Task" [ 831.554343] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.564720] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041055, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.600791] env[61964]: DEBUG nova.compute.manager [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 831.600981] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.602027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4655b370-553e-4d7b-8208-46ff4751540a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.611149] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.611376] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b990ee96-e4de-45d5-8f71-c43494f4af02 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.621732] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 831.621732] env[61964]: value = "task-1041056" [ 831.621732] env[61964]: _type = "Task" [ 831.621732] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.643630] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.643922] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041046, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.653889] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 831.787778] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096177} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.791363] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.792406] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32139aa-bc3e-4888-99c6-0c1ec69e0d5b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.813615] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041052, 'name': Rename_Task, 'duration_secs': 0.201058} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.822889] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.823346] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 831.823577] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fc71f3b-7077-4b51-92da-42dd791f52a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.840477] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c671207-0a34-4849-af8d-825d2522d1a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.850574] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 831.850574] env[61964]: value = "task-1041058" [ 831.850574] env[61964]: _type = "Task" [ 831.850574] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.852416] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 831.852416] env[61964]: value = "task-1041057" [ 831.852416] env[61964]: _type = "Task" [ 831.852416] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.866168] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041057, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.870065] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041058, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.879373] env[61964]: DEBUG nova.compute.manager [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 831.879688] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.880158] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05a0b4f4-8a60-4a9c-9b81-7f583e64b7e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.890421] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 831.890421] env[61964]: value = "task-1041059" [ 831.890421] env[61964]: _type = "Task" [ 831.890421] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.900703] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.013942] env[61964]: DEBUG oslo_vmware.api [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041050, 'name': ReconfigVM_Task, 'duration_secs': 0.783035} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.014515] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.014747] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfigured VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 832.034671] env[61964]: DEBUG oslo_concurrency.lockutils [None req-93053d75-d3ae-49fe-8e24-8f31aa6992c5 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "062da016-f4d6-424d-90e6-5903940e2daa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.537s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.063119] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updated VIF entry in instance network info cache for port df79c16f-2899-499e-ab0f-f8f7752913e5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 832.063529] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [{"id": "df79c16f-2899-499e-ab0f-f8f7752913e5", "address": "fa:16:3e:83:c8:1d", "network": {"id": "a25357b6-1448-4eff-a07f-3b464d6aecb7", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1941905063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19bac13e505143b4ae5884c24c748b1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "418ddd3d-5f64-407e-8e0c-c8b81639bee9", "external-id": "nsx-vlan-transportzone-107", "segmentation_id": 107, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf79c16f-28", "ovs_interfaceid": "df79c16f-2899-499e-ab0f-f8f7752913e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.071301] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041055, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.136628] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041056, 'name': PowerOffVM_Task, 'duration_secs': 0.257153} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.143032] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.143032] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 832.143346] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041046, 'name': CloneVM_Task, 'duration_secs': 1.680278} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.144072] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fd89437-74f1-40ce-9452-2f2e294436e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.146106] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Created linked-clone VM from snapshot [ 832.147421] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06dec30-6be0-4954-be0d-6c3d225b8d21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.168182] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Uploading image acef627a-6665-4737-891e-74c2350a5188 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 832.185277] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.200584] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 832.200584] env[61964]: value = "vm-230492" [ 832.200584] env[61964]: _type = "VirtualMachine" [ 832.200584] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 832.200991] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b017dadd-6e2b-4174-848a-93cd512ed0fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.214462] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease: (returnval){ [ 832.214462] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3be14-2a00-27fe-0d95-8107f05b713f" [ 832.214462] env[61964]: _type = "HttpNfcLease" [ 832.214462] env[61964]: } obtained for exporting VM: (result){ [ 832.214462] env[61964]: value = "vm-230492" [ 832.214462] env[61964]: _type = "VirtualMachine" [ 832.214462] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 832.214851] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the lease: (returnval){ [ 832.214851] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3be14-2a00-27fe-0d95-8107f05b713f" [ 832.214851] env[61964]: _type = "HttpNfcLease" [ 832.214851] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 832.230100] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 832.230100] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3be14-2a00-27fe-0d95-8107f05b713f" [ 832.230100] env[61964]: _type = "HttpNfcLease" [ 832.230100] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 832.236415] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 832.236921] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 832.237175] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleting the datastore file [datastore2] 2a185e4b-c636-418a-a75e-2cf158f550c0 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.237467] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffbe9983-18bb-4c9c-81ed-4c0cd2c5c32b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.248322] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 832.248322] env[61964]: value = "task-1041062" [ 832.248322] env[61964]: _type = "Task" [ 832.248322] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.260830] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041062, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.371498] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041058, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.375948] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041057, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.405632] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041059, 'name': PowerOffVM_Task, 'duration_secs': 0.429574} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.406090] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.406417] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 832.406743] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230481', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'name': 'volume-a9253389-533d-4822-88cf-2d7417a75289', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c922228-4ba9-43ed-8e5c-3ca15e48ceab', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'serial': 'a9253389-533d-4822-88cf-2d7417a75289'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 832.407898] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17afa10c-3aac-4f46-8323-0110453297c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.437789] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790aca30-3b13-4418-921c-e63ef06ebb3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.441950] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e12dd3-28f5-4a4d-ba54-b884945a3e2c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.454680] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446886d0-a25b-450d-8e6a-90fa7de5b89e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.459515] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cf249e-2168-4c44-bb7c-d6692f4d0904 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.522512] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223da3c4-3ca9-4bc2-8404-249a944a8619 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.526525] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a24bc063-dc74-4e18-a30e-0cdbb8579953 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.342s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.528811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb52b45-59e4-41fc-b91b-dabc4c8e8acb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.548218] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] The volume has not been displaced from its original location: [datastore1] volume-a9253389-533d-4822-88cf-2d7417a75289/volume-a9253389-533d-4822-88cf-2d7417a75289.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 832.554864] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 832.558302] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d682b159-a354-4ca9-b927-1528e2345198 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.573492] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Releasing lock "refresh_cache-2a185e4b-c636-418a-a75e-2cf158f550c0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.573840] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-plugged-15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 832.574117] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.574379] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.574601] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.574838] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] No waiting events found dispatching network-vif-plugged-15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 832.575077] env[61964]: WARNING nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received unexpected event network-vif-plugged-15153388-0bbd-4e56-b179-42a3c5ad9552 for instance with vm_state active and task_state None. [ 832.575336] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-changed-15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 832.575521] env[61964]: DEBUG nova.compute.manager [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing instance network info cache due to event network-changed-15153388-0bbd-4e56-b179-42a3c5ad9552. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 832.575718] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.575863] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.576040] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing network info cache for port 15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.581843] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df3a010-7dfc-4233-98a1-5f5f7ede16df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.604353] env[61964]: DEBUG nova.compute.provider_tree [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.605606] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041055, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.607374] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 832.607374] env[61964]: value = "task-1041063" [ 832.607374] env[61964]: _type = "Task" [ 832.607374] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.618685] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.724103] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 832.724103] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3be14-2a00-27fe-0d95-8107f05b713f" [ 832.724103] env[61964]: _type = "HttpNfcLease" [ 832.724103] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 832.724485] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 832.724485] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3be14-2a00-27fe-0d95-8107f05b713f" [ 832.724485] env[61964]: _type = "HttpNfcLease" [ 832.724485] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 832.725248] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f40c8a-e736-4bc9-93cc-baa1b2b51208 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.733907] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 832.734126] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 832.816150] env[61964]: DEBUG oslo_vmware.api [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041062, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228132} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.816429] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.816625] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 832.816806] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 832.817049] env[61964]: INFO nova.compute.manager [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Took 1.22 seconds to destroy the instance on the hypervisor. [ 832.817348] env[61964]: DEBUG oslo.service.loopingcall [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.817553] env[61964]: DEBUG nova.compute.manager [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 832.817687] env[61964]: DEBUG nova.network.neutron [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 832.863750] env[61964]: DEBUG oslo_vmware.api [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041058, 'name': PowerOnVM_Task, 'duration_secs': 0.719453} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.870840] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 832.871255] env[61964]: INFO nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Took 8.89 seconds to spawn the instance on the hypervisor. [ 832.871518] env[61964]: DEBUG nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.873453] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9fffba-06ab-4755-ad56-f827a7c509cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.876787] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d2296e31-31e7-449a-bcfe-380efefc542f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.893467] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041057, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.094820] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041055, 'name': CreateVM_Task, 'duration_secs': 1.429588} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.095572] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 833.096445] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.096645] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.097501] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.098128] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88c6cead-61d8-4027-a9a3-e819f6220a3f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.106076] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 833.106076] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e48560-6092-2bba-267d-e138ac2cd032" [ 833.106076] env[61964]: _type = "Task" [ 833.106076] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.107448] env[61964]: DEBUG nova.scheduler.client.report [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 833.130762] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041063, 'name': ReconfigVM_Task, 'duration_secs': 0.352844} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.136026] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 833.143142] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e48560-6092-2bba-267d-e138ac2cd032, 'name': SearchDatastore_Task, 'duration_secs': 0.014542} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.143142] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-842ea754-678d-4733-95ec-916b85fe1ef1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.153321] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.153590] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.153843] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.153996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.154199] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.154862] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18ed5fef-5e0f-4a96-873e-8fef302e4f82 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.170075] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updated VIF entry in instance network info cache for port 15153388-0bbd-4e56-b179-42a3c5ad9552. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.171040] env[61964]: DEBUG nova.network.neutron [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15153388-0bbd-4e56-b179-42a3c5ad9552", "address": "fa:16:3e:86:91:41", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15153388-0b", "ovs_interfaceid": "15153388-0bbd-4e56-b179-42a3c5ad9552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.176174] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.176318] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 833.177238] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 833.177238] env[61964]: value = "task-1041064" [ 833.177238] env[61964]: _type = "Task" [ 833.177238] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.180113] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bea06b2a-cb78-4616-b83f-043a3588d61a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.195485] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041064, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.197305] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 833.197305] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbb7ab-c4c2-4ce3-2a0e-04f1566435a0" [ 833.197305] env[61964]: _type = "Task" [ 833.197305] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.211152] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbb7ab-c4c2-4ce3-2a0e-04f1566435a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.376779] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041057, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.404695] env[61964]: INFO nova.compute.manager [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Took 29.82 seconds to build instance. [ 833.554029] env[61964]: DEBUG nova.compute.manager [req-638b533c-6e99-45fc-b345-4390c8f95350 req-94eed75c-f193-4aeb-a855-b0f17f29d9b9 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Received event network-vif-deleted-df79c16f-2899-499e-ab0f-f8f7752913e5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 833.554674] env[61964]: INFO nova.compute.manager [req-638b533c-6e99-45fc-b345-4390c8f95350 req-94eed75c-f193-4aeb-a855-b0f17f29d9b9 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Neutron deleted interface df79c16f-2899-499e-ab0f-f8f7752913e5; detaching it from the instance and deleting it from the info cache [ 833.554829] env[61964]: DEBUG nova.network.neutron [req-638b533c-6e99-45fc-b345-4390c8f95350 req-94eed75c-f193-4aeb-a855-b0f17f29d9b9 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.601107] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 833.602288] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5d398a-a918-410a-8e75-f91b1b8ce67d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.616309] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 833.616309] env[61964]: ERROR oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk due to incomplete transfer. [ 833.616309] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4d02e82c-2fbf-414e-9d29-2f46898067c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.619574] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.125s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.622990] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.284s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.623776] env[61964]: DEBUG nova.objects.instance [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lazy-loading 'resources' on Instance uuid cdfc0bf1-d603-4580-8527-8b06e5ae0799 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.633301] env[61964]: DEBUG nova.compute.manager [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Received event network-changed-3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 833.634017] env[61964]: DEBUG nova.compute.manager [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Refreshing instance network info cache due to event network-changed-3a986bcd-6584-4440-bf01-7db0dff75b2f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 833.635010] env[61964]: DEBUG oslo_concurrency.lockutils [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] Acquiring lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.635010] env[61964]: DEBUG oslo_concurrency.lockutils [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] Acquired lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.635010] env[61964]: DEBUG nova.network.neutron [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Refreshing network info cache for port 3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.640804] env[61964]: DEBUG oslo_vmware.rw_handles [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52007df2-a98c-b445-7b92-276aea677560/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 833.640896] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Uploaded image f04b1f74-105c-4cda-9f12-0aea28fe5e43 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 833.644515] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 833.648203] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5c685686-97b4-4852-8d2b-e4ca15cfdefc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.659325] env[61964]: INFO nova.scheduler.client.report [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Deleted allocations for instance 560bd013-eae1-4541-aaa2-0732d8f20e4a [ 833.669239] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 833.669239] env[61964]: value = "task-1041065" [ 833.669239] env[61964]: _type = "Task" [ 833.669239] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.683953] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0b41f95-91d0-417f-ad3c-5a0bdb2b85a8 req-0823cd9f-3f6e-4df6-97ff-086545749f7f service nova] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.684247] env[61964]: DEBUG nova.network.neutron [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.686096] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.699627] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041064, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.717343] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbb7ab-c4c2-4ce3-2a0e-04f1566435a0, 'name': SearchDatastore_Task, 'duration_secs': 0.017733} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.718406] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38e4c786-9917-4831-902e-e8a648b3e883 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.726484] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 833.726484] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d8f58b-d579-24c4-91bb-d394ac892d8b" [ 833.726484] env[61964]: _type = "Task" [ 833.726484] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.744123] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d8f58b-d579-24c4-91bb-d394ac892d8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.877906] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041057, 'name': ReconfigVM_Task, 'duration_secs': 1.567866} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.878262] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Reconfigured VM instance instance-00000046 to attach disk [datastore1] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.879189] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b259c9b-0a78-4bb9-951b-1ccc390242f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.891027] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 833.891027] env[61964]: value = "task-1041066" [ 833.891027] env[61964]: _type = "Task" [ 833.891027] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.901665] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041066, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.906641] env[61964]: DEBUG oslo_concurrency.lockutils [None req-643578b8-74a5-4d5b-a4a3-acdc8ab7b7e8 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.339s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.059267] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5f83286-aaa6-4697-9e44-15d09ef73620 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.070759] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431d28df-5f7c-467b-8aaf-7426b50cb09f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.114792] env[61964]: DEBUG nova.compute.manager [req-638b533c-6e99-45fc-b345-4390c8f95350 req-94eed75c-f193-4aeb-a855-b0f17f29d9b9 service nova] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Detach interface failed, port_id=df79c16f-2899-499e-ab0f-f8f7752913e5, reason: Instance 2a185e4b-c636-418a-a75e-2cf158f550c0 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 834.184858] env[61964]: DEBUG oslo_concurrency.lockutils [None req-573fcc10-21e1-4e7c-8acd-8ed7c365b012 tempest-MultipleCreateTestJSON-719907479 tempest-MultipleCreateTestJSON-719907479-project-member] Lock "560bd013-eae1-4541-aaa2-0732d8f20e4a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.936s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.191778] env[61964]: INFO nova.compute.manager [-] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Took 1.37 seconds to deallocate network for instance. [ 834.199201] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.208793] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041064, 'name': ReconfigVM_Task, 'duration_secs': 0.954231} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.212287] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230481', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'name': 'volume-a9253389-533d-4822-88cf-2d7417a75289', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c922228-4ba9-43ed-8e5c-3ca15e48ceab', 'attached_at': '', 'detached_at': '', 'volume_id': 'a9253389-533d-4822-88cf-2d7417a75289', 'serial': 'a9253389-533d-4822-88cf-2d7417a75289'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 834.213116] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 834.217060] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b922de5f-464a-4cdd-bc20-80983de364bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.228289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.234669] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e26aaf27-a024-4394-879a-89a985456397 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.245137] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d8f58b-d579-24c4-91bb-d394ac892d8b, 'name': SearchDatastore_Task, 'duration_secs': 0.021194} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.245933] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.246360] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 834.247690] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aedbe771-a17e-4cb4-9721-f8db6f040db7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.263293] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 834.263293] env[61964]: value = "task-1041068" [ 834.263293] env[61964]: _type = "Task" [ 834.263293] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.274770] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041068, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.335557] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.335795] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.335985] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore2] 1c922228-4ba9-43ed-8e5c-3ca15e48ceab {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.336311] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85c63dd7-dc65-4c86-8744-3d7174ccbbbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.344568] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 834.344568] env[61964]: value = "task-1041069" [ 834.344568] env[61964]: _type = "Task" [ 834.344568] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.357981] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041069, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.407442] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041066, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.522084] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0df482-365d-44d4-b1d8-ec0851cd41aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.531900] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50bf9ab-947d-4783-b4a1-1e839ad7787c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.569026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb66adb6-c1b9-4db2-9106-f54ffe9ca84a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.579132] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadb91e1-9967-4fdf-a2f5-b85866190db1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.601526] env[61964]: DEBUG nova.compute.provider_tree [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.686012] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.708869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.785403] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041068, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.852892] env[61964]: DEBUG nova.network.neutron [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updated VIF entry in instance network info cache for port 3a986bcd-6584-4440-bf01-7db0dff75b2f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.853452] env[61964]: DEBUG nova.network.neutron [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.865310] env[61964]: DEBUG oslo_vmware.api [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041069, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366463} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.865598] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.866715] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.866715] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.866715] env[61964]: INFO nova.compute.manager [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Took 2.99 seconds to destroy the instance on the hypervisor. [ 834.866715] env[61964]: DEBUG oslo.service.loopingcall [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.866715] env[61964]: DEBUG nova.compute.manager [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 834.866715] env[61964]: DEBUG nova.network.neutron [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.902683] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-fdf326b7-a14c-40cb-bfbb-569b3af8fb77" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.902942] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-fdf326b7-a14c-40cb-bfbb-569b3af8fb77" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.903348] env[61964]: DEBUG nova.objects.instance [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'flavor' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.904709] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041066, 'name': Rename_Task, 'duration_secs': 0.663498} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.907771] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.907771] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-082138c6-294e-4f11-9655-84ea2f5c4af1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.913985] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 834.913985] env[61964]: value = "task-1041070" [ 834.913985] env[61964]: _type = "Task" [ 834.913985] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.924735] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041070, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.103767] env[61964]: DEBUG nova.scheduler.client.report [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.186624] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.279924] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041068, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572851} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.280244] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 835.280429] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.280704] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b29f84de-f492-472a-836c-3499d3d27143 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.294019] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 835.294019] env[61964]: value = "task-1041071" [ 835.294019] env[61964]: _type = "Task" [ 835.294019] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.303554] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.361083] env[61964]: DEBUG oslo_concurrency.lockutils [req-7a06b66c-05ef-4b40-b58f-c0f77be7e8ca req-4c0c1985-9450-48cd-b8b3-6ca1bad5d100 service nova] Releasing lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.426136] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041070, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.567256] env[61964]: DEBUG nova.objects.instance [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'pci_requests' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.610572] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.986s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.611313] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.201s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.611777] env[61964]: DEBUG nova.objects.instance [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lazy-loading 'resources' on Instance uuid beddafaf-8a91-4bfa-af3c-fa69977677ba {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.686276] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.691940] env[61964]: DEBUG nova.network.neutron [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.801597] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.911928] env[61964]: DEBUG nova.compute.manager [req-e6ebaaf5-327c-4f5f-88ec-d8b91247b824 req-f679f4dd-10c4-4396-b2ae-4dd1442d965c service nova] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Received event network-vif-deleted-9d295b40-9665-4671-9ea8-108dbb9169d7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 835.927318] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041070, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.068827] env[61964]: DEBUG nova.objects.base [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 836.069166] env[61964]: DEBUG nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 836.134612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d7fbfebd-7960-4688-bdff-28fefcadc27e tempest-ServerActionsV293TestJSON-1482128244 tempest-ServerActionsV293TestJSON-1482128244-project-member] Lock "cdfc0bf1-d603-4580-8527-8b06e5ae0799" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.711s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.190481] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041065, 'name': Destroy_Task, 'duration_secs': 2.236939} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.193410] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroyed the VM [ 836.193684] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 836.194093] env[61964]: INFO nova.compute.manager [-] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Took 1.33 seconds to deallocate network for instance. [ 836.195638] env[61964]: DEBUG nova.policy [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.197435] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-832efbee-43cf-48c4-810c-1b2c0eb1f89e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.208766] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 836.208766] env[61964]: value = "task-1041072" [ 836.208766] env[61964]: _type = "Task" [ 836.208766] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.221849] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041072, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.308967] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041071, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.702903} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.309499] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.310578] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8d0e20-93e5-4dbb-b8b4-2bcf79eb8c69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.333136] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.336452] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a751a8ca-8182-449d-9027-86891bdb3467 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.359370] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 836.359370] env[61964]: value = "task-1041073" [ 836.359370] env[61964]: _type = "Task" [ 836.359370] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.368234] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041073, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.431339] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041070, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.540070] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422ed19f-94f0-410a-b6e0-30d0cbdb2d46 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.553218] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe637056-ef92-4c5d-a987-459d7b38a53e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.594112] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f287ef-ea9a-4db3-b923-7d2b00c3142e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.604044] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d028ea11-3f05-4385-a77d-7e1ac71d2a17 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.620355] env[61964]: DEBUG nova.compute.provider_tree [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.724611] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041072, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.759569] env[61964]: INFO nova.compute.manager [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Took 0.56 seconds to detach 1 volumes for instance. [ 836.874027] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.926258] env[61964]: DEBUG oslo_vmware.api [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041070, 'name': PowerOnVM_Task, 'duration_secs': 1.584433} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.926563] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 836.926750] env[61964]: INFO nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Took 8.51 seconds to spawn the instance on the hypervisor. [ 836.926935] env[61964]: DEBUG nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 836.927978] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f959c878-3402-416c-9cd4-d80d9a7b419f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.125163] env[61964]: DEBUG nova.scheduler.client.report [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 837.223545] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041072, 'name': RemoveSnapshot_Task} progress is 61%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.269055] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.277482] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.277768] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.277987] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.278191] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.278373] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.280874] env[61964]: INFO nova.compute.manager [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Terminating instance [ 837.370057] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041073, 'name': ReconfigVM_Task, 'duration_secs': 0.573512} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.370361] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.370997] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-709b1658-d7eb-40e2-b2bf-58874388a96e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.378952] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 837.378952] env[61964]: value = "task-1041074" [ 837.378952] env[61964]: _type = "Task" [ 837.378952] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.387525] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041074, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.446529] env[61964]: INFO nova.compute.manager [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Took 28.04 seconds to build instance. [ 837.631030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.020s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.636068] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.179s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.637684] env[61964]: INFO nova.compute.claims [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.678241] env[61964]: INFO nova.scheduler.client.report [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleted allocations for instance beddafaf-8a91-4bfa-af3c-fa69977677ba [ 837.727581] env[61964]: DEBUG oslo_vmware.api [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041072, 'name': RemoveSnapshot_Task, 'duration_secs': 1.069889} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.728569] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 837.728907] env[61964]: INFO nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 16.48 seconds to snapshot the instance on the hypervisor. [ 837.784389] env[61964]: DEBUG nova.compute.manager [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 837.784712] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 837.786102] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600bbf57-3775-40cd-bb7f-4a74377eb024 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.795669] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.795948] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4e0dfef-13b4-4f0a-a926-5499b91fc957 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.803705] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 837.803705] env[61964]: value = "task-1041075" [ 837.803705] env[61964]: _type = "Task" [ 837.803705] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.813437] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.890654] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041074, 'name': Rename_Task, 'duration_secs': 0.281611} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.890958] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 837.891239] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-328c73cb-0821-4f16-a882-c449211a51f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.899765] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 837.899765] env[61964]: value = "task-1041076" [ 837.899765] env[61964]: _type = "Task" [ 837.899765] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.910109] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041076, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.948423] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49a31447-dff4-43a1-9037-6b985f46bff1 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.561s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.957861] env[61964]: DEBUG nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Successfully updated port: fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.115447] env[61964]: DEBUG nova.compute.manager [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-plugged-fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 838.115681] env[61964]: DEBUG oslo_concurrency.lockutils [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.115903] env[61964]: DEBUG oslo_concurrency.lockutils [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.116090] env[61964]: DEBUG oslo_concurrency.lockutils [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.116268] env[61964]: DEBUG nova.compute.manager [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] No waiting events found dispatching network-vif-plugged-fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 838.116461] env[61964]: WARNING nova.compute.manager [req-8fc7fa67-82e1-42c8-b5d3-f8627ffac55a req-cdae1547-d093-4189-af05-7d8508dac2a7 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received unexpected event network-vif-plugged-fdf326b7-a14c-40cb-bfbb-569b3af8fb77 for instance with vm_state active and task_state None. [ 838.191027] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e5c7ba37-f95c-470f-b3ea-65fbae6ede4f tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "beddafaf-8a91-4bfa-af3c-fa69977677ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.379s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.283033] env[61964]: DEBUG nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Found 3 images (rotation: 2) {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4985}} [ 838.283388] env[61964]: DEBUG nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Rotating out 1 backups {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4993}} [ 838.283480] env[61964]: DEBUG nova.compute.manager [None req-daf40723-5433-4822-8ebb-eecf8a5afde2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleting image a4bc1cdf-2405-42f5-a7fb-757fd7700645 {{(pid=61964) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4998}} [ 838.314978] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041075, 'name': PowerOffVM_Task, 'duration_secs': 0.300206} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.315308] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 838.315550] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 838.315834] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc002632-628c-4b84-a5c3-6750b39718fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.410523] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041076, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.462031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.462031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.462031] env[61964]: DEBUG nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.911241] env[61964]: DEBUG oslo_vmware.api [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041076, 'name': PowerOnVM_Task, 'duration_secs': 0.838579} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.911544] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 838.911728] env[61964]: INFO nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Took 7.49 seconds to spawn the instance on the hypervisor. [ 838.911938] env[61964]: DEBUG nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 838.912961] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369e28ef-22f5-4425-abc1-6001184cf236 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.951334] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7323303-f36a-4752-8a17-7d2d6b947a73 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.960841] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b982b5-3abe-4326-a91f-001664d4d126 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.995772] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55fe81f-fa8a-4edb-93b1-d8ddae20afdf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.004742] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2ccf42-8703-4d66-9fb4-704e71bdfe14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.027052] env[61964]: DEBUG nova.compute.provider_tree [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.030315] env[61964]: WARNING nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] 48dd64e7-3659-4104-8273-6f58f8d3eb44 already exists in list: networks containing: ['48dd64e7-3659-4104-8273-6f58f8d3eb44']. ignoring it [ 839.031379] env[61964]: WARNING nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] 48dd64e7-3659-4104-8273-6f58f8d3eb44 already exists in list: networks containing: ['48dd64e7-3659-4104-8273-6f58f8d3eb44']. ignoring it [ 839.123220] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "16276be9-b305-4d1d-afde-bc98be42687a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.123496] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.123713] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "16276be9-b305-4d1d-afde-bc98be42687a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.123900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.124085] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.128555] env[61964]: INFO nova.compute.manager [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Terminating instance [ 839.279534] env[61964]: INFO nova.compute.manager [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Rebuilding instance [ 839.327221] env[61964]: DEBUG nova.compute.manager [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 839.328113] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7140a6-ddec-4f6f-8a7b-ec80a9b5a9c2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.434918] env[61964]: INFO nova.compute.manager [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Took 27.61 seconds to build instance. [ 839.534587] env[61964]: DEBUG nova.scheduler.client.report [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 839.634844] env[61964]: DEBUG nova.compute.manager [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 839.635110] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.636169] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db100ca5-c571-430a-96a7-e9a58fb58f0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.646488] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.646810] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2ed1f5b-a136-4041-9c62-5fbe2e71ac0a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.655163] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 839.655163] env[61964]: value = "task-1041078" [ 839.655163] env[61964]: _type = "Task" [ 839.655163] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.666330] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.676741] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.676913] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.677275] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleting the datastore file [datastore1] b90e7ebd-8dc6-495b-99bf-15ff24763af4 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.677551] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-abdf4484-c978-48c9-88bf-17f3fa9266be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.683122] env[61964]: DEBUG nova.network.neutron [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15153388-0bbd-4e56-b179-42a3c5ad9552", "address": "fa:16:3e:86:91:41", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15153388-0b", "ovs_interfaceid": "15153388-0bbd-4e56-b179-42a3c5ad9552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "address": "fa:16:3e:83:58:2b", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdf326b7-a1", "ovs_interfaceid": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.692365] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 839.692365] env[61964]: value = "task-1041079" [ 839.692365] env[61964]: _type = "Task" [ 839.692365] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.703862] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.937131] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba9d6440-43ec-41c4-9791-103932086314 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.161s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.042026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.042687] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 840.045969] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.776s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.046887] env[61964]: INFO nova.compute.claims [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.165923] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041078, 'name': PowerOffVM_Task, 'duration_secs': 0.280322} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.166396] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.166592] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.166856] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f04cec6-385e-4a7d-b842-e2b3c653e148 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.186453] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.187460] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.187614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.188819] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411b5306-1730-4324-821b-13a6a508a774 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.204368] env[61964]: DEBUG oslo_vmware.api [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162751} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.218316] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.218615] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.218830] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.219028] env[61964]: INFO nova.compute.manager [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Took 2.43 seconds to destroy the instance on the hypervisor. [ 840.219323] env[61964]: DEBUG oslo.service.loopingcall [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.220402] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 840.220707] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.220948] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 840.221682] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.221682] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 840.221682] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 840.221852] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 840.221888] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 840.223104] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 840.223104] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 840.223104] env[61964]: DEBUG nova.virt.hardware [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 840.229665] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfiguring VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 840.230078] env[61964]: DEBUG nova.compute.manager [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 840.230189] env[61964]: DEBUG nova.network.neutron [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.231983] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c2be8e2-720d-439f-aea4-109a39419489 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.248474] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.248634] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.248811] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleting the datastore file [datastore2] 16276be9-b305-4d1d-afde-bc98be42687a {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.249281] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd373646-8614-4b6d-b101-77f2eb1d9969 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.257656] env[61964]: DEBUG oslo_vmware.api [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 840.257656] env[61964]: value = "task-1041081" [ 840.257656] env[61964]: _type = "Task" [ 840.257656] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.257656] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for the task: (returnval){ [ 840.257656] env[61964]: value = "task-1041082" [ 840.257656] env[61964]: _type = "Task" [ 840.257656] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.273301] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041082, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.273566] env[61964]: DEBUG oslo_vmware.api [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041081, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.345600] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.348470] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92e9b951-c927-46a2-89af-90e70ae4db07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.360240] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 840.360240] env[61964]: value = "task-1041083" [ 840.360240] env[61964]: _type = "Task" [ 840.360240] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.370294] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.551701] env[61964]: DEBUG nova.compute.utils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.558812] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 840.558991] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 840.576617] env[61964]: DEBUG nova.compute.manager [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-changed-fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 840.576871] env[61964]: DEBUG nova.compute.manager [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing instance network info cache due to event network-changed-fdf326b7-a14c-40cb-bfbb-569b3af8fb77. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 840.577629] env[61964]: DEBUG oslo_concurrency.lockutils [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.577629] env[61964]: DEBUG oslo_concurrency.lockutils [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.577629] env[61964]: DEBUG nova.network.neutron [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Refreshing network info cache for port fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.602538] env[61964]: DEBUG nova.policy [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1cb6721924d64cbc96bc8d4c8d08d681', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '618ce6bbecae4f50af04142b697d2828', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.771741] env[61964]: DEBUG oslo_vmware.api [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Task: {'id': task-1041082, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245405} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.775381] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.775582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.775700] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.775885] env[61964]: INFO nova.compute.manager [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 840.776161] env[61964]: DEBUG oslo.service.loopingcall [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.776405] env[61964]: DEBUG oslo_vmware.api [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041081, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.776600] env[61964]: DEBUG nova.compute.manager [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 840.776693] env[61964]: DEBUG nova.network.neutron [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.856705] env[61964]: INFO nova.compute.manager [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Rebuilding instance [ 840.860390] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Successfully created port: 388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.871518] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041083, 'name': PowerOffVM_Task, 'duration_secs': 0.195532} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.873959] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.874341] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 840.875629] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc139af-bf92-4b91-b30b-edd94455795a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.886381] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.886764] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3aebc59d-8338-4514-8ece-ff0e65b05fa7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.913427] env[61964]: DEBUG nova.compute.manager [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 840.914324] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da89290-b4de-4b44-a3d8-db1d8e88475e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.918169] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.918386] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.918583] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Deleting the datastore file [datastore1] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.919285] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a18ba8a2-1c03-461a-9e8e-6b7f9f445fbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.926855] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 840.926855] env[61964]: value = "task-1041085" [ 840.926855] env[61964]: _type = "Task" [ 840.926855] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.938033] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.062021] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 841.065882] env[61964]: DEBUG nova.network.neutron [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.276617] env[61964]: DEBUG oslo_vmware.api [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041081, 'name': ReconfigVM_Task, 'duration_secs': 0.606476} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.276617] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.276617] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfigured VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 841.396584] env[61964]: DEBUG nova.network.neutron [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updated VIF entry in instance network info cache for port fdf326b7-a14c-40cb-bfbb-569b3af8fb77. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 841.397156] env[61964]: DEBUG nova.network.neutron [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "15153388-0bbd-4e56-b179-42a3c5ad9552", "address": "fa:16:3e:86:91:41", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15153388-0b", "ovs_interfaceid": "15153388-0bbd-4e56-b179-42a3c5ad9552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "address": "fa:16:3e:83:58:2b", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdf326b7-a1", "ovs_interfaceid": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.414410] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b79695-7727-429c-9db1-04bac0f021e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.423134] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04dcd42-5450-4a38-8869-236fb4778b0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.459897] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 841.460707] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4931d3-e4fd-4896-946c-2d95cd2d3e4f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.463491] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d1bd2c5-909e-4bef-89a8-a209ae872d71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.476761] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3000c2-2a6e-47a6-bc53-0607df66d5c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.480878] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 841.480878] env[61964]: value = "task-1041086" [ 841.480878] env[61964]: _type = "Task" [ 841.480878] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.481284] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194713} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.481811] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.482037] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 841.482228] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 841.496553] env[61964]: DEBUG nova.compute.provider_tree [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.504242] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.543754] env[61964]: DEBUG nova.network.neutron [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.571951] env[61964]: INFO nova.compute.manager [-] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Took 1.34 seconds to deallocate network for instance. [ 841.780534] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd4a0a9-6544-40f0-ab1a-4048a5317836 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-fdf326b7-a14c-40cb-bfbb-569b3af8fb77" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.877s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.899831] env[61964]: DEBUG oslo_concurrency.lockutils [req-648fe728-960e-4ba9-90fa-c58ef3b985f0 req-7e21a3e2-3734-47aa-a3ea-4603273f5ca2 service nova] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.991988] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041086, 'name': PowerOffVM_Task, 'duration_secs': 0.14972} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.992301] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 841.992543] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 841.993656] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a110ec3a-b2dd-4d7d-af04-122b0573a2f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.001135] env[61964]: DEBUG nova.scheduler.client.report [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.010453] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 842.010763] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a557b32-d124-4a02-8739-5f07dfdd8abb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.042223] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 842.042510] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 842.042761] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Deleting the datastore file [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.043045] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be8803f4-e184-4c02-8b9c-f05a87a42c99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.046091] env[61964]: INFO nova.compute.manager [-] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Took 1.27 seconds to deallocate network for instance. [ 842.054047] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 842.054047] env[61964]: value = "task-1041088" [ 842.054047] env[61964]: _type = "Task" [ 842.054047] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.062530] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.074298] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 842.078473] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.097890] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 842.098187] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.098345] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 842.098564] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.098794] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 842.098968] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 842.099221] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 842.099422] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 842.099613] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 842.099794] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 842.099983] env[61964]: DEBUG nova.virt.hardware [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 842.101269] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6605edd-c85f-4a59-8b8f-59a70c66f428 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.111815] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0c268a-1871-4040-8b72-0be1d3e786ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.418294] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Successfully updated port: 388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.508883] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.509464] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 842.512889] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.303s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.513145] env[61964]: DEBUG nova.objects.instance [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lazy-loading 'resources' on Instance uuid ab33fb75-4114-4a26-8692-befb407b7460 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.536154] env[61964]: DEBUG nova.compute.manager [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Received event network-vif-plugged-388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 842.536552] env[61964]: DEBUG oslo_concurrency.lockutils [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] Acquiring lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.536552] env[61964]: DEBUG oslo_concurrency.lockutils [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.536647] env[61964]: DEBUG oslo_concurrency.lockutils [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.536880] env[61964]: DEBUG nova.compute.manager [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] No waiting events found dispatching network-vif-plugged-388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 842.537138] env[61964]: WARNING nova.compute.manager [req-a133eb1b-be56-44e1-85b5-f0e604773a95 req-f6dbc2d0-f638-4359-85ce-ebd1b71bc922 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Received unexpected event network-vif-plugged-388f0e3c-60fd-49b3-b81f-5637bae949ca for instance with vm_state building and task_state spawning. [ 842.539827] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 842.540915] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.540915] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 842.540915] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.540915] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 842.540915] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 842.541535] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 842.542162] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 842.542162] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 842.542324] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 842.542471] env[61964]: DEBUG nova.virt.hardware [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 842.544218] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bae701-2eea-44cd-812e-37b4f9f9222c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.556139] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.561745] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37396e20-0be0-463a-86bf-aef596771a0c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.578932] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.586084] env[61964]: DEBUG oslo.service.loopingcall [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.590340] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 842.590340] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206815} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.590340] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09cb42c5-782b-4eb4-9f2a-7c31a08f7501 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.601939] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.602134] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.602323] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.612296] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.612296] env[61964]: value = "task-1041089" [ 842.612296] env[61964]: _type = "Task" [ 842.612296] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.626668] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041089, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.751485] env[61964]: DEBUG nova.compute.manager [req-3d6bd6e7-0abe-408d-97d3-c6e59c25364d req-fce327b4-2efc-4634-864c-7a87079d0b73 service nova] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Received event network-vif-deleted-19290ca0-b730-490e-b3a8-5a598373bcc7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 842.751718] env[61964]: DEBUG nova.compute.manager [req-3d6bd6e7-0abe-408d-97d3-c6e59c25364d req-fce327b4-2efc-4634-864c-7a87079d0b73 service nova] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Received event network-vif-deleted-41feaa65-ae5c-49a5-828a-4e85d7aecb93 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 842.789185] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 842.790151] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be946095-9f0a-4892-9cfd-c0623cb88049 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.797250] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 842.797439] env[61964]: ERROR oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk due to incomplete transfer. [ 842.797732] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-432b884f-77e9-44fb-aa38-932aa206b4be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.806941] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5230c8f3-72c9-4859-47cb-aa8394c795c6/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 842.807335] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Uploaded image acef627a-6665-4737-891e-74c2350a5188 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 842.809230] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 842.809521] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-33a69fd7-eb0e-41d7-9c11-06e5253a3f8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.818376] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 842.818376] env[61964]: value = "task-1041090" [ 842.818376] env[61964]: _type = "Task" [ 842.818376] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.828940] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041090, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.921110] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.921289] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquired lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.921470] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.017568] env[61964]: DEBUG nova.compute.utils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.020063] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 843.020282] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 843.073668] env[61964]: DEBUG nova.policy [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.123554] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041089, 'name': CreateVM_Task, 'duration_secs': 0.335337} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.123777] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 843.124225] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.124403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.124761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.127498] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e431d86-e62c-4ade-b89f-f9a3acd3fe40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.133953] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 843.133953] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52878a08-3ffb-50e2-2a4e-fce927158243" [ 843.133953] env[61964]: _type = "Task" [ 843.133953] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.147919] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52878a08-3ffb-50e2-2a4e-fce927158243, 'name': SearchDatastore_Task, 'duration_secs': 0.011004} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.148911] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.149196] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.149474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.149633] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.149818] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.150369] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6df5887-251f-4fbb-8050-958f3abd43bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.163328] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.163539] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.164298] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44dcaea6-0218-48e6-a38a-2972e4e1e96c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.176588] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 843.176588] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52908e06-3ac3-ccee-f465-6b785a9377e6" [ 843.176588] env[61964]: _type = "Task" [ 843.176588] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.187837] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52908e06-3ac3-ccee-f465-6b785a9377e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.331448] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041090, 'name': Destroy_Task, 'duration_secs': 0.3807} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.332730] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Destroyed the VM [ 843.332981] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 843.333731] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb4a43d-4999-46b9-9c6e-afb09d33341e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.336811] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b123e9c7-4936-4b2d-9f38-b26fdf672aef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.343978] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a12c82e-487e-4aac-aeb2-17303172215f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.349532] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 843.349532] env[61964]: value = "task-1041091" [ 843.349532] env[61964]: _type = "Task" [ 843.349532] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.378748] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3c2e5e-f758-40f9-b256-8f3562976b45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.384537] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041091, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.390626] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab68ee73-7f56-4a2c-bb4c-469613df2364 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.408236] env[61964]: DEBUG nova.compute.provider_tree [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.461894] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.475294] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Successfully created port: 6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.523842] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 843.640140] env[61964]: DEBUG nova.network.neutron [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Updating instance_info_cache with network_info: [{"id": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "address": "fa:16:3e:b7:0a:f3", "network": {"id": "782dcd4e-e1e1-48ff-9fd0-94b2673b0183", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-386045169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "618ce6bbecae4f50af04142b697d2828", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap388f0e3c-60", "ovs_interfaceid": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.652649] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 843.653015] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.654541] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 843.654541] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.654541] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 843.654541] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 843.654730] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 843.654927] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 843.655193] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 843.655445] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 843.656788] env[61964]: DEBUG nova.virt.hardware [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 843.656788] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea457ff-4409-420c-b94a-1853bd904a08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.667957] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a77ae4-5f1b-4233-ae15-a30d8e8c40d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.688154] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.696475] env[61964]: DEBUG oslo.service.loopingcall [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.702026] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.702026] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f7eed3f-3bb7-4b25-b652-da7f5c670148 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.719013] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-15153388-0bbd-4e56-b179-42a3c5ad9552" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.719289] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-15153388-0bbd-4e56-b179-42a3c5ad9552" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.720661] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52908e06-3ac3-ccee-f465-6b785a9377e6, 'name': SearchDatastore_Task, 'duration_secs': 0.010629} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.723226] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.723226] env[61964]: value = "task-1041092" [ 843.723226] env[61964]: _type = "Task" [ 843.723226] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.723647] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7397239e-8ce8-468f-bfd5-b911d58d2423 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.733658] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 843.733658] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bc8a50-684f-48db-c656-9c383db30454" [ 843.733658] env[61964]: _type = "Task" [ 843.733658] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.742401] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041092, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.754617] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bc8a50-684f-48db-c656-9c383db30454, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.809246] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.809403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.809638] env[61964]: DEBUG nova.compute.manager [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.810540] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017a44bf-8b13-4395-af4c-63cfee0c6066 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.819112] env[61964]: DEBUG nova.compute.manager [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 843.819854] env[61964]: DEBUG nova.objects.instance [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.861587] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041091, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.911707] env[61964]: DEBUG nova.scheduler.client.report [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 844.143415] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Releasing lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.143782] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Instance network_info: |[{"id": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "address": "fa:16:3e:b7:0a:f3", "network": {"id": "782dcd4e-e1e1-48ff-9fd0-94b2673b0183", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-386045169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "618ce6bbecae4f50af04142b697d2828", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap388f0e3c-60", "ovs_interfaceid": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 844.144276] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:0a:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f54f7284-8f7d-47ee-839d-2143062cfe44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '388f0e3c-60fd-49b3-b81f-5637bae949ca', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.152052] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Creating folder: Project (618ce6bbecae4f50af04142b697d2828). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 844.152503] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8cdd2853-b1c1-4d05-8e19-ca0456cd7e08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.167310] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Created folder: Project (618ce6bbecae4f50af04142b697d2828) in parent group-v230360. [ 844.168036] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Creating folder: Instances. Parent ref: group-v230498. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 844.168036] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c406f8ef-ba5f-40f9-8d4d-e12dc8f26456 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.178490] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Created folder: Instances in parent group-v230498. [ 844.178760] env[61964]: DEBUG oslo.service.loopingcall [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.178990] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 844.179708] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bdff51a1-6dd4-4bde-926e-a30b6c33de3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.200429] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.200429] env[61964]: value = "task-1041095" [ 844.200429] env[61964]: _type = "Task" [ 844.200429] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.208599] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041095, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.224449] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.224688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.225588] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cef7de1-ead4-4b0c-8692-70d83e3cb34f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.239177] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041092, 'name': CreateVM_Task, 'duration_secs': 0.346114} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.251445] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.252971] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.252971] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.252971] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.256506] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ef446e-2fd3-4a55-9f81-e805cd088897 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.259264] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dc663cb-4ec2-479f-85ab-52c5109045b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.267169] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bc8a50-684f-48db-c656-9c383db30454, 'name': SearchDatastore_Task, 'duration_secs': 0.03198} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.287837] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.287837] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.287953] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 844.287953] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c9a108-5dc1-cd6b-6bbf-69fc8330310a" [ 844.287953] env[61964]: _type = "Task" [ 844.287953] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.294643] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfiguring VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 844.294985] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ae1b812-25cf-4e12-bb49-19844086f8dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.297248] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eb10b42-96d9-48ba-ae5c-f1a71e6ffe23 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.320080] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 844.320080] env[61964]: value = "task-1041096" [ 844.320080] env[61964]: _type = "Task" [ 844.320080] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.325830] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c9a108-5dc1-cd6b-6bbf-69fc8330310a, 'name': SearchDatastore_Task, 'duration_secs': 0.012518} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.331326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.331326] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.331326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.331326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.332850] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.332850] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 844.332850] env[61964]: value = "task-1041097" [ 844.332850] env[61964]: _type = "Task" [ 844.332850] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.332850] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6911c5a-2282-4c58-a351-8f969a2294d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.342486] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041096, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.349780] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.359522] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.359786] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.363480] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc0d2549-a38d-4394-8691-f91674289bb4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.365900] env[61964]: DEBUG oslo_vmware.api [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041091, 'name': RemoveSnapshot_Task, 'duration_secs': 0.693959} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.366469] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 844.366759] env[61964]: INFO nova.compute.manager [None req-2e9b3a27-f9a3-497d-9b1b-34c96f36401d tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Took 16.46 seconds to snapshot the instance on the hypervisor. [ 844.373381] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 844.373381] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5d4c7-666c-7189-6ebc-343f29433bfe" [ 844.373381] env[61964]: _type = "Task" [ 844.373381] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.383665] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5d4c7-666c-7189-6ebc-343f29433bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.416747] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.419863] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.754s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.419863] env[61964]: DEBUG nova.objects.instance [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 844.434689] env[61964]: INFO nova.scheduler.client.report [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleted allocations for instance ab33fb75-4114-4a26-8692-befb407b7460 [ 844.538022] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 844.564511] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 844.564781] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.564946] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 844.565235] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.565356] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 844.565448] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 844.565670] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 844.565836] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 844.566014] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 844.566211] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 844.566391] env[61964]: DEBUG nova.virt.hardware [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 844.567390] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8620496e-0f95-4b20-b646-15d94794c6b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.576336] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e18abb-76ef-4707-89f5-bb4f0e5ffc15 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.627857] env[61964]: DEBUG nova.compute.manager [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Received event network-changed-388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 844.628210] env[61964]: DEBUG nova.compute.manager [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Refreshing instance network info cache due to event network-changed-388f0e3c-60fd-49b3-b81f-5637bae949ca. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 844.628396] env[61964]: DEBUG oslo_concurrency.lockutils [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] Acquiring lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.628560] env[61964]: DEBUG oslo_concurrency.lockutils [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] Acquired lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.628792] env[61964]: DEBUG nova.network.neutron [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Refreshing network info cache for port 388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.712789] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041095, 'name': CreateVM_Task, 'duration_secs': 0.453464} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.712789] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.713029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.713217] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.713626] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.713969] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78955238-b690-4503-a77f-edf954f404cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.721271] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 844.721271] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52df1ada-8177-134b-17b2-3e22d3551b49" [ 844.721271] env[61964]: _type = "Task" [ 844.721271] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.733360] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52df1ada-8177-134b-17b2-3e22d3551b49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.814839] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "50b04cbf-9064-4010-b350-638e8096bb96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.815090] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.838046] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 844.838912] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041096, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.839211] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4abd5549-f817-4ad7-aaf9-542f9f17dc4c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.850891] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.852314] env[61964]: DEBUG oslo_vmware.api [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 844.852314] env[61964]: value = "task-1041098" [ 844.852314] env[61964]: _type = "Task" [ 844.852314] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.862534] env[61964]: DEBUG oslo_vmware.api [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.887071] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b5d4c7-666c-7189-6ebc-343f29433bfe, 'name': SearchDatastore_Task, 'duration_secs': 0.012635} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.888665] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-043eab3b-39bd-44ee-bfd1-f5953db49566 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.895644] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 844.895644] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46c3f-07dd-11db-128b-49539707ca6e" [ 844.895644] env[61964]: _type = "Task" [ 844.895644] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.904821] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46c3f-07dd-11db-128b-49539707ca6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.943305] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d38cb345-10e1-42dd-917f-ba32a1678e63 tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.161s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.944603] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 19.605s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.947469] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "ab33fb75-4114-4a26-8692-befb407b7460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.947469] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.947469] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.948932] env[61964]: INFO nova.compute.manager [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Terminating instance [ 845.211460] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Successfully updated port: 6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.234524] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52df1ada-8177-134b-17b2-3e22d3551b49, 'name': SearchDatastore_Task, 'duration_secs': 0.055002} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.234847] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.235096] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.235316] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.318116] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 845.352229] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041096, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541306} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.353276] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.353821] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.354516] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e223714-923b-49ca-b991-8b0103bd7470 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.367883] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.372621] env[61964]: DEBUG oslo_vmware.api [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041098, 'name': PowerOffVM_Task, 'duration_secs': 0.216361} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.373304] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 845.373304] env[61964]: value = "task-1041099" [ 845.373304] env[61964]: _type = "Task" [ 845.373304] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.373645] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.373875] env[61964]: DEBUG nova.compute.manager [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 845.380030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5578d4-71d8-41f4-86ce-ae924f331525 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.391664] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.407749] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46c3f-07dd-11db-128b-49539707ca6e, 'name': SearchDatastore_Task, 'duration_secs': 0.010617} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.408086] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.408381] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.408701] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.408905] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.409167] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-422e4edf-b5d6-4cd1-a941-79ba9ddb9893 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.411273] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60a9537c-27cb-4e51-ad61-d2acb52ba877 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.420224] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 845.420224] env[61964]: value = "task-1041100" [ 845.420224] env[61964]: _type = "Task" [ 845.420224] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.431205] env[61964]: DEBUG oslo_concurrency.lockutils [None req-62e75a23-8ed0-49d5-b792-0afca8080d5d tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.432322] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.435371] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.061s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.437427] env[61964]: INFO nova.compute.claims [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.449140] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.449364] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.451260] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d758466-964d-4369-a4bf-5780b94d531d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.454390] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.454597] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquired lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.454840] env[61964]: DEBUG nova.network.neutron [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.461563] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 845.461563] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284924e-1013-1729-b6df-90a390b3c6b5" [ 845.461563] env[61964]: _type = "Task" [ 845.461563] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.474420] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284924e-1013-1729-b6df-90a390b3c6b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.512012] env[61964]: DEBUG nova.network.neutron [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Updated VIF entry in instance network info cache for port 388f0e3c-60fd-49b3-b81f-5637bae949ca. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 845.512384] env[61964]: DEBUG nova.network.neutron [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Updating instance_info_cache with network_info: [{"id": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "address": "fa:16:3e:b7:0a:f3", "network": {"id": "782dcd4e-e1e1-48ff-9fd0-94b2673b0183", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-386045169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "618ce6bbecae4f50af04142b697d2828", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f54f7284-8f7d-47ee-839d-2143062cfe44", "external-id": "nsx-vlan-transportzone-569", "segmentation_id": 569, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap388f0e3c-60", "ovs_interfaceid": "388f0e3c-60fd-49b3-b81f-5637bae949ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.715314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.715543] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.715651] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.853387] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.860904] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.887738] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066856} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.888482] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.889603] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a655aab-d14f-439b-96e7-74adaf40a4c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.913921] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.914648] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1f6235f5-6ffd-4ceb-a501-995e08208ddd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.105s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.915732] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40d421a2-5184-4c45-b518-c38eb5353ae2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.952939] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041100, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.955344] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 845.955344] env[61964]: value = "task-1041101" [ 845.955344] env[61964]: _type = "Task" [ 845.955344] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.959392] env[61964]: DEBUG nova.compute.utils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Can not refresh info_cache because instance was not found {{(pid=61964) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1055}} [ 845.971107] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041101, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.979280] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284924e-1013-1729-b6df-90a390b3c6b5, 'name': SearchDatastore_Task, 'duration_secs': 0.043899} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.980506] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da16580e-d2ee-40c1-af64-e09b2629b9af {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.983931] env[61964]: DEBUG nova.network.neutron [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.991182] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 845.991182] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f12ba4-1c86-feb4-3cd5-3e351b301d76" [ 845.991182] env[61964]: _type = "Task" [ 845.991182] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.000815] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f12ba4-1c86-feb4-3cd5-3e351b301d76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.015769] env[61964]: DEBUG oslo_concurrency.lockutils [req-764518a1-dfc4-4fdf-b02a-24d674bcbd4f req-c1f1f611-3ff4-43a2-a346-bd12d146a216 service nova] Releasing lock "refresh_cache-eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.115299] env[61964]: DEBUG nova.network.neutron [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.267352] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.357940] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.444775] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "46435c68-f85a-4360-b2b5-6296afc33c3e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.445118] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.445393] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.445606] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.445834] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.451943] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75265} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.453753] env[61964]: INFO nova.compute.manager [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Terminating instance [ 846.454618] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.454890] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.463932] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8252ac4-1717-4a03-8f6d-1231b4cb254d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.473351] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041101, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.476039] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 846.476039] env[61964]: value = "task-1041102" [ 846.476039] env[61964]: _type = "Task" [ 846.476039] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.503983] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f12ba4-1c86-feb4-3cd5-3e351b301d76, 'name': SearchDatastore_Task, 'duration_secs': 0.061707} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.504191] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.504465] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] eb97402e-e722-4cc3-a1d3-1b360ab1e1a7/eb97402e-e722-4cc3-a1d3-1b360ab1e1a7.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 846.504982] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7a8102d-e3dc-4b61-bfa0-0ae6157a6a25 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.517579] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 846.517579] env[61964]: value = "task-1041103" [ 846.517579] env[61964]: _type = "Task" [ 846.517579] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.527107] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.618600] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Releasing lock "refresh_cache-ab33fb75-4114-4a26-8692-befb407b7460" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.619138] env[61964]: DEBUG nova.compute.manager [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 846.619355] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 846.619681] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7811ec0a-4ca2-415d-91b5-0f4ee13a2d42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.640832] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fd200e-bf14-4519-8550-9b90ff444429 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.677677] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab33fb75-4114-4a26-8692-befb407b7460 could not be found. [ 846.678087] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.678164] env[61964]: INFO nova.compute.manager [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 0.06 seconds to destroy the instance on the hypervisor. [ 846.678423] env[61964]: DEBUG oslo.service.loopingcall [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.681170] env[61964]: DEBUG nova.compute.manager [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 846.681393] env[61964]: DEBUG nova.network.neutron [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.701086] env[61964]: DEBUG nova.network.neutron [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.710799] env[61964]: DEBUG nova.compute.manager [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Received event network-vif-plugged-6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 846.711145] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Acquiring lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.711287] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.711438] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.711614] env[61964]: DEBUG nova.compute.manager [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] No waiting events found dispatching network-vif-plugged-6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 846.711785] env[61964]: WARNING nova.compute.manager [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Received unexpected event network-vif-plugged-6467ee6e-8b51-47b3-8441-d53dd716c656 for instance with vm_state building and task_state spawning. [ 846.712067] env[61964]: DEBUG nova.compute.manager [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Received event network-changed-6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 846.712268] env[61964]: DEBUG nova.compute.manager [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Refreshing instance network info cache due to event network-changed-6467ee6e-8b51-47b3-8441-d53dd716c656. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 846.712447] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Acquiring lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.752856] env[61964]: DEBUG nova.network.neutron [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Updating instance_info_cache with network_info: [{"id": "6467ee6e-8b51-47b3-8441-d53dd716c656", "address": "fa:16:3e:a4:a4:00", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6467ee6e-8b", "ovs_interfaceid": "6467ee6e-8b51-47b3-8441-d53dd716c656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.817986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fc6167-5172-4df8-978d-a8160cff7100 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.827358] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ce76f3-3ae4-47b2-b106-a9260920e6ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.867658] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafb3777-0f5b-4cb1-a34a-0d1f8512444a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.879732] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3d08b9-beff-4869-a761-1ce473b0aba4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.884052] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.897178] env[61964]: DEBUG nova.compute.provider_tree [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.963530] env[61964]: DEBUG nova.compute.manager [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 846.963530] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 846.967383] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9a3dcc-6969-4fd8-8a73-4ff88be2a03d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.978134] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041101, 'name': ReconfigVM_Task, 'duration_secs': 0.701608} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.978217] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Reconfigured VM instance instance-00000046 to attach disk [datastore2] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0/e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.978935] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 846.979156] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-beaede92-51ba-4e2a-b6cf-0ec6762c6f6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.984931] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd24847c-4eaf-4bd2-a303-5244ca849ccf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.992666] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085425} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.994930] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.995297] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 846.995297] env[61964]: value = "task-1041105" [ 846.995297] env[61964]: _type = "Task" [ 846.995297] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.995604] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 846.995604] env[61964]: value = "task-1041104" [ 846.995604] env[61964]: _type = "Task" [ 846.995604] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.996327] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c06877-6d61-4d65-9dd9-1269d3f1ef5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.029063] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.029495] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041104, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.033364] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c38dfe59-f8eb-445b-86a3-6b4451a09170 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.052174] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.059540] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469252} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.060919] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] eb97402e-e722-4cc3-a1d3-1b360ab1e1a7/eb97402e-e722-4cc3-a1d3-1b360ab1e1a7.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.061225] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.061562] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 847.061562] env[61964]: value = "task-1041106" [ 847.061562] env[61964]: _type = "Task" [ 847.061562] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.061819] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4aa9895-a70d-47e6-823d-578c976baca9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.073021] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041106, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.074715] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 847.074715] env[61964]: value = "task-1041107" [ 847.074715] env[61964]: _type = "Task" [ 847.074715] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.083307] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.206807] env[61964]: DEBUG nova.network.neutron [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.255368] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.255807] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Instance network_info: |[{"id": "6467ee6e-8b51-47b3-8441-d53dd716c656", "address": "fa:16:3e:a4:a4:00", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6467ee6e-8b", "ovs_interfaceid": "6467ee6e-8b51-47b3-8441-d53dd716c656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 847.256227] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Acquired lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.256480] env[61964]: DEBUG nova.network.neutron [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Refreshing network info cache for port 6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.257835] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:a4:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6467ee6e-8b51-47b3-8441-d53dd716c656', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.265932] env[61964]: DEBUG oslo.service.loopingcall [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.267044] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 847.267323] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d4bca0d-302b-48dc-9ee1-6f908a0bb052 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.289709] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.289709] env[61964]: value = "task-1041108" [ 847.289709] env[61964]: _type = "Task" [ 847.289709] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.298716] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041108, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.374088] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.400768] env[61964]: DEBUG nova.scheduler.client.report [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.511918] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041105, 'name': PowerOffVM_Task, 'duration_secs': 0.230671} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.515027] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 847.515226] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 847.515512] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041104, 'name': Rename_Task, 'duration_secs': 0.21071} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.515717] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9c29717-9d35-4da3-8ed6-870cfffb26cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.517875] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.517875] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24c8b6bd-f44c-4e76-a9ff-874fa73a136b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.526764] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 847.526764] env[61964]: value = "task-1041109" [ 847.526764] env[61964]: _type = "Task" [ 847.526764] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.535549] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.575483] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041106, 'name': ReconfigVM_Task, 'duration_secs': 0.29919} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.575631] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688/93720ce5-834a-4d8d-b230-13d38620b688.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.579149] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-029c16de-fb1c-46de-8f2b-2fe807a7ae69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.588947] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07892} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.590569] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.590992] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 847.590992] env[61964]: value = "task-1041111" [ 847.590992] env[61964]: _type = "Task" [ 847.590992] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.591742] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b40ecd-d88e-4bc2-a7a4-ae1a7f52c68b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.600095] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 847.600326] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 847.600531] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleting the datastore file [datastore2] 46435c68-f85a-4360-b2b5-6296afc33c3e {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.600792] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ec5618d-bb49-48b9-b515-f20dd8c2143f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.606533] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041111, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.617306] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for the task: (returnval){ [ 847.617306] env[61964]: value = "task-1041112" [ 847.617306] env[61964]: _type = "Task" [ 847.617306] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.626695] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] eb97402e-e722-4cc3-a1d3-1b360ab1e1a7/eb97402e-e722-4cc3-a1d3-1b360ab1e1a7.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.627116] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96c1e0ed-3cee-4975-bdf4-8a0b319bbf64 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.654037] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.654037] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 847.654037] env[61964]: value = "task-1041113" [ 847.654037] env[61964]: _type = "Task" [ 847.654037] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.661425] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.709442] env[61964]: INFO nova.compute.manager [-] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Took 1.03 seconds to deallocate network for instance. [ 847.800287] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041108, 'name': CreateVM_Task, 'duration_secs': 0.40194} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.800580] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 847.801430] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.801642] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.802063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.803700] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fa5b488-eb02-44cc-937a-7485ac85f307 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.810019] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 847.810019] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52267a4e-aa34-eba5-0751-4e085cdb13da" [ 847.810019] env[61964]: _type = "Task" [ 847.810019] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.819218] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52267a4e-aa34-eba5-0751-4e085cdb13da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.822319] env[61964]: DEBUG nova.compute.manager [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Stashing vm_state: stopped {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 847.876807] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.907156] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.907156] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 847.909329] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.412s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.909987] env[61964]: DEBUG nova.objects.instance [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'resources' on Instance uuid a23e7f40-1b3e-4c8f-b664-30f6314923a1 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.040643] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041109, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.101970] env[61964]: DEBUG nova.network.neutron [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Updated VIF entry in instance network info cache for port 6467ee6e-8b51-47b3-8441-d53dd716c656. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.102588] env[61964]: DEBUG nova.network.neutron [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Updating instance_info_cache with network_info: [{"id": "6467ee6e-8b51-47b3-8441-d53dd716c656", "address": "fa:16:3e:a4:a4:00", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6467ee6e-8b", "ovs_interfaceid": "6467ee6e-8b51-47b3-8441-d53dd716c656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.112761] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041111, 'name': Rename_Task, 'duration_secs': 0.163596} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.113313] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.113604] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-901cdefc-ca4e-4af6-8638-fde1dac76cbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.123148] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 848.123148] env[61964]: value = "task-1041114" [ 848.123148] env[61964]: _type = "Task" [ 848.123148] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.140783] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041114, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.145705] env[61964]: DEBUG oslo_vmware.api [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Task: {'id': task-1041112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181138} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.147029] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.147029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 848.147029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 848.147029] env[61964]: INFO nova.compute.manager [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 848.147414] env[61964]: DEBUG oslo.service.loopingcall [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.147718] env[61964]: DEBUG nova.compute.manager [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 848.147863] env[61964]: DEBUG nova.network.neutron [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 848.168783] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041113, 'name': ReconfigVM_Task, 'duration_secs': 0.304855} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.169230] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Reconfigured VM instance instance-00000048 to attach disk [datastore2] eb97402e-e722-4cc3-a1d3-1b360ab1e1a7/eb97402e-e722-4cc3-a1d3-1b360ab1e1a7.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.170180] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f3e9148-4000-4e89-aa44-f0275232491c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.179156] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 848.179156] env[61964]: value = "task-1041115" [ 848.179156] env[61964]: _type = "Task" [ 848.179156] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.192070] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041115, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.216536] env[61964]: INFO nova.compute.manager [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance disappeared during terminate [ 848.216789] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3c81d201-7e21-4041-a7d1-e2d9d86b905f tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "ab33fb75-4114-4a26-8692-befb407b7460" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.272s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.320628] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52267a4e-aa34-eba5-0751-4e085cdb13da, 'name': SearchDatastore_Task, 'duration_secs': 0.012203} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.320990] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.321241] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.321499] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.321626] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.321806] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.322119] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5784aec-6f62-494e-a6a5-b3f5eb35051e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.343054] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.343284] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 848.344853] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.345474] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5372fbab-c99b-4674-8693-cc9ac4694ec3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.351951] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 848.351951] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5250f503-f3b4-2b0a-d873-7c8633eee280" [ 848.351951] env[61964]: _type = "Task" [ 848.351951] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.362974] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5250f503-f3b4-2b0a-d873-7c8633eee280, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.378768] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.413097] env[61964]: DEBUG nova.compute.utils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.413097] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 848.413323] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.515705] env[61964]: DEBUG nova.policy [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f438e739578a4fabad898ebd5e2925cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ccf1fe5003e484fa73ce5c32767479f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.539567] env[61964]: DEBUG oslo_vmware.api [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041109, 'name': PowerOnVM_Task, 'duration_secs': 0.518046} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.544604] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.544604] env[61964]: DEBUG nova.compute.manager [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 848.547121] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53446361-be6c-41ec-baca-14a828444807 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.573580] env[61964]: DEBUG nova.compute.manager [req-fa73b162-0ef2-4d11-8e25-a4399e6965bb req-d6ba723d-c0b2-4fab-8bd0-23007964901b service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Received event network-vif-deleted-89204ea4-fd9b-4979-8aaf-f224046531fa {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 848.573782] env[61964]: INFO nova.compute.manager [req-fa73b162-0ef2-4d11-8e25-a4399e6965bb req-d6ba723d-c0b2-4fab-8bd0-23007964901b service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Neutron deleted interface 89204ea4-fd9b-4979-8aaf-f224046531fa; detaching it from the instance and deleting it from the info cache [ 848.573955] env[61964]: DEBUG nova.network.neutron [req-fa73b162-0ef2-4d11-8e25-a4399e6965bb req-d6ba723d-c0b2-4fab-8bd0-23007964901b service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.607072] env[61964]: DEBUG oslo_concurrency.lockutils [req-4dbe1639-d596-487d-81b3-ac12ff73900d req-832f9111-894d-4779-8dae-89a895fea1e1 service nova] Releasing lock "refresh_cache-fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.637823] env[61964]: DEBUG oslo_vmware.api [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041114, 'name': PowerOnVM_Task, 'duration_secs': 0.446928} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.638139] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.638349] env[61964]: DEBUG nova.compute.manager [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 848.639213] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab1f5f8-5146-4090-952f-068c11d59ec0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.691946] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041115, 'name': Rename_Task, 'duration_secs': 0.196199} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.692280] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.692532] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad3b50a2-3af9-4f43-b47a-2367384bb720 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.701113] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 848.701113] env[61964]: value = "task-1041116" [ 848.701113] env[61964]: _type = "Task" [ 848.701113] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.709516] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.771109] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f94ffe-a66e-4fb0-820f-62332bc851e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.780764] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d9598b-527f-404e-9dd4-29d1de054c6f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.818439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f6ed8b-1ce9-4819-93a7-469678fb87d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.828528] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6eb42ad-3c69-4ef0-8a2d-697cc6c16b46 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.844375] env[61964]: DEBUG nova.compute.provider_tree [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.865224] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5250f503-f3b4-2b0a-d873-7c8633eee280, 'name': SearchDatastore_Task, 'duration_secs': 0.013487} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.870969] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5600209-a681-41ff-8a99-44e4da7bef75 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.877545] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 848.877545] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ee86-1bca-e8c9-10a8-94609684b8fb" [ 848.877545] env[61964]: _type = "Task" [ 848.877545] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.880955] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.890211] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ee86-1bca-e8c9-10a8-94609684b8fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.910564] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Successfully created port: 2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.921597] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 848.968410] env[61964]: DEBUG nova.network.neutron [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.071157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.076905] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3160c840-c9f7-48e4-a006-64b81fc03e05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.087479] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b97fda-2dba-49df-8cd0-e05bac28cc1b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.125480] env[61964]: DEBUG nova.compute.manager [req-fa73b162-0ef2-4d11-8e25-a4399e6965bb req-d6ba723d-c0b2-4fab-8bd0-23007964901b service nova] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Detach interface failed, port_id=89204ea4-fd9b-4979-8aaf-f224046531fa, reason: Instance 46435c68-f85a-4360-b2b5-6296afc33c3e could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 849.160526] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.211330] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041116, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.347188] env[61964]: DEBUG nova.scheduler.client.report [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.378490] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.392994] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b8ee86-1bca-e8c9-10a8-94609684b8fb, 'name': SearchDatastore_Task, 'duration_secs': 0.018912} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.393571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.393681] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] fb96bf3a-ea2c-414d-bb29-eca2b07d51b6/fb96bf3a-ea2c-414d-bb29-eca2b07d51b6.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.394889] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0cc9bab-c79a-44eb-9256-590586297495 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.405575] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 849.405575] env[61964]: value = "task-1041117" [ 849.405575] env[61964]: _type = "Task" [ 849.405575] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.413558] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.471148] env[61964]: INFO nova.compute.manager [-] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Took 1.32 seconds to deallocate network for instance. [ 849.691239] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "93720ce5-834a-4d8d-b230-13d38620b688" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.691484] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.691697] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "93720ce5-834a-4d8d-b230-13d38620b688-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.691905] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.692097] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.697973] env[61964]: INFO nova.compute.manager [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Terminating instance [ 849.713941] env[61964]: DEBUG oslo_vmware.api [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041116, 'name': PowerOnVM_Task, 'duration_secs': 0.868596} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.715398] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.715767] env[61964]: INFO nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Took 7.64 seconds to spawn the instance on the hypervisor. [ 849.716489] env[61964]: DEBUG nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.717571] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4191cc-ae83-44b8-91bb-60e9b80c8b5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.853755] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.855607] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.670s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.856852] env[61964]: INFO nova.compute.claims [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.879202] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.880422] env[61964]: INFO nova.scheduler.client.report [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocations for instance a23e7f40-1b3e-4c8f-b664-30f6314923a1 [ 849.916289] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041117, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.932872] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 849.964281] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 849.964543] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.964711] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 849.964899] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.965060] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 849.965310] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 849.965578] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 849.965743] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 849.965910] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 849.966084] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 849.966260] env[61964]: DEBUG nova.virt.hardware [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 849.967224] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da8cc24-d256-45d8-b950-d5c5b2bddfbc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.976909] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e7f0b2-73f1-4094-a50c-349e4de7df7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.981818] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.202217] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.202516] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.202739] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.202932] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.203120] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.205376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "refresh_cache-93720ce5-834a-4d8d-b230-13d38620b688" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.205620] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquired lock "refresh_cache-93720ce5-834a-4d8d-b230-13d38620b688" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.205725] env[61964]: DEBUG nova.network.neutron [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.206936] env[61964]: INFO nova.compute.manager [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Terminating instance [ 850.235246] env[61964]: INFO nova.compute.manager [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Took 32.81 seconds to build instance. [ 850.309687] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.309870] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.309934] env[61964]: INFO nova.compute.manager [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Shelving [ 850.379825] env[61964]: DEBUG oslo_vmware.api [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041097, 'name': ReconfigVM_Task, 'duration_secs': 5.870312} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.380464] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.380721] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Reconfigured VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 850.387823] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e546f76-b638-4f67-ae0b-0c7238bdb91f tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "a23e7f40-1b3e-4c8f-b664-30f6314923a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.495s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.415561] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58897} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.415829] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] fb96bf3a-ea2c-414d-bb29-eca2b07d51b6/fb96bf3a-ea2c-414d-bb29-eca2b07d51b6.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 850.416058] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.416322] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be1f3949-501f-42b5-ba95-9a9b62400280 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.424269] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 850.424269] env[61964]: value = "task-1041118" [ 850.424269] env[61964]: _type = "Task" [ 850.424269] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.432984] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041118, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.528858] env[61964]: DEBUG nova.compute.manager [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Received event network-vif-plugged-2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 850.529128] env[61964]: DEBUG oslo_concurrency.lockutils [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.529353] env[61964]: DEBUG oslo_concurrency.lockutils [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.529531] env[61964]: DEBUG oslo_concurrency.lockutils [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.529741] env[61964]: DEBUG nova.compute.manager [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] No waiting events found dispatching network-vif-plugged-2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 850.529913] env[61964]: WARNING nova.compute.manager [req-00d6e701-0a6c-4a3c-9bc7-4e3945dd46b0 req-0d9ab9e2-cd06-4b86-a905-615b344887b9 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Received unexpected event network-vif-plugged-2a884ac5-7364-4168-bf0a-d947f84b8b92 for instance with vm_state building and task_state spawning. [ 850.711921] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "refresh_cache-e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.712159] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquired lock "refresh_cache-e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.712341] env[61964]: DEBUG nova.network.neutron [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.725325] env[61964]: DEBUG nova.network.neutron [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.736905] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d341f7c8-a4e4-407b-a2ea-660283ba0dee tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.314s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.738319] env[61964]: DEBUG nova.compute.manager [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-deleted-15153388-0bbd-4e56-b179-42a3c5ad9552 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 850.738319] env[61964]: INFO nova.compute.manager [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Neutron deleted interface 15153388-0bbd-4e56-b179-42a3c5ad9552; detaching it from the instance and deleting it from the info cache [ 850.738618] env[61964]: DEBUG nova.network.neutron [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "address": "fa:16:3e:83:58:2b", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdf326b7-a1", "ovs_interfaceid": "fdf326b7-a14c-40cb-bfbb-569b3af8fb77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.781028] env[61964]: DEBUG nova.network.neutron [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.939140] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041118, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072753} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.939140] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.939398] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e0d8e8-5c7f-4163-919b-86c3f3fbe1ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.968460] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] fb96bf3a-ea2c-414d-bb29-eca2b07d51b6/fb96bf3a-ea2c-414d-bb29-eca2b07d51b6.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.971817] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e475e8b4-78b0-4169-b3e0-78c0569f9899 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.993802] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 850.993802] env[61964]: value = "task-1041119" [ 850.993802] env[61964]: _type = "Task" [ 850.993802] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.004430] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.004430] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Successfully updated port: 2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.126723] env[61964]: DEBUG nova.compute.manager [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Received event network-changed-2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 851.126979] env[61964]: DEBUG nova.compute.manager [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Refreshing instance network info cache due to event network-changed-2a884ac5-7364-4168-bf0a-d947f84b8b92. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 851.127235] env[61964]: DEBUG oslo_concurrency.lockutils [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] Acquiring lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.127384] env[61964]: DEBUG oslo_concurrency.lockutils [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] Acquired lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.127565] env[61964]: DEBUG nova.network.neutron [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Refreshing network info cache for port 2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.230177] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8029860-b6b4-4db8-9a30-f3999fe51dbd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.239028] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd84ebd2-9efb-4917-9f58-7cb924ff1459 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.244159] env[61964]: DEBUG nova.network.neutron [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.246324] env[61964]: DEBUG oslo_concurrency.lockutils [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.246503] env[61964]: DEBUG oslo_concurrency.lockutils [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] Acquired lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.277641] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed5ff01-66fa-438d-8492-aa73f051e8e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.282425] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1bb2b2-e7bd-472f-8fdb-30674926c4d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.285756] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Releasing lock "refresh_cache-93720ce5-834a-4d8d-b230-13d38620b688" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.286193] env[61964]: DEBUG nova.compute.manager [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 851.286417] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 851.287217] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7bb366-7842-43e9-93ef-3c8922676ba0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.303972] env[61964]: DEBUG oslo_concurrency.lockutils [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] Releasing lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.304368] env[61964]: WARNING nova.compute.manager [req-5f3b5cdd-850a-4ce3-ba6b-88c099e286f4 req-cf2361cc-9c0e-4444-9140-dc6757508cab service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Detach interface failed, port_id=15153388-0bbd-4e56-b179-42a3c5ad9552, reason: No device with interface-id 15153388-0bbd-4e56-b179-42a3c5ad9552 exists on VM: nova.exception.NotFound: No device with interface-id 15153388-0bbd-4e56-b179-42a3c5ad9552 exists on VM [ 851.313027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34e11aa-eaa4-464a-9ee8-e0b4e5deefd5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.322148] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.322148] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3197cceb-e95a-4a09-b0b5-e2e7c8bcd481 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.322913] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.331970] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b7b79a5-9761-4e20-abcf-f4d87ea8b435 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.334250] env[61964]: DEBUG nova.compute.provider_tree [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.339195] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 851.339195] env[61964]: value = "task-1041120" [ 851.339195] env[61964]: _type = "Task" [ 851.339195] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.343835] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 851.343835] env[61964]: value = "task-1041121" [ 851.343835] env[61964]: _type = "Task" [ 851.343835] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.353106] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.358700] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.376359] env[61964]: DEBUG nova.network.neutron [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.506125] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041119, 'name': ReconfigVM_Task, 'duration_secs': 0.340486} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.506125] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Reconfigured VM instance instance-00000049 to attach disk [datastore2] fb96bf3a-ea2c-414d-bb29-eca2b07d51b6/fb96bf3a-ea2c-414d-bb29-eca2b07d51b6.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.506125] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aae9b1bf-79c8-4b88-8d7a-f707642df19a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.511901] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.514879] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 851.514879] env[61964]: value = "task-1041122" [ 851.514879] env[61964]: _type = "Task" [ 851.514879] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.523892] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041122, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.661415] env[61964]: DEBUG nova.network.neutron [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.680636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.680829] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.681026] env[61964]: DEBUG nova.network.neutron [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.739395] env[61964]: DEBUG nova.network.neutron [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.838812] env[61964]: DEBUG nova.scheduler.client.report [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.855508] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041120, 'name': PowerOffVM_Task, 'duration_secs': 0.218404} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.860241] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.860441] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 851.861133] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041121, 'name': PowerOffVM_Task, 'duration_secs': 0.420162} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.861490] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7fb51e2c-5eb8-4257-8e8d-010fbc7563b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.863265] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.864283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c212c4fc-9f36-4ba9-86fd-7fa79187bc66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.883031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Releasing lock "refresh_cache-e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.883532] env[61964]: DEBUG nova.compute.manager [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 851.883689] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 851.884756] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7725f7f9-90c9-4942-8ce2-5fca8b63a18e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.887994] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90725a97-c963-4a56-bd15-e8c70b3addd9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.893043] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 851.893263] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 851.893450] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Deleting the datastore file [datastore2] 93720ce5-834a-4d8d-b230-13d38620b688 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.893688] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c488a4f-207d-4878-9dd3-c90ebf263702 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.898790] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.902065] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2dd751ff-4417-40e7-b3c4-f95234cbec45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.906208] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for the task: (returnval){ [ 851.906208] env[61964]: value = "task-1041124" [ 851.906208] env[61964]: _type = "Task" [ 851.906208] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.910555] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 851.910555] env[61964]: value = "task-1041125" [ 851.910555] env[61964]: _type = "Task" [ 851.910555] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.916958] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.917137] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.917347] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.918009] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.918009] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.919468] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.920319] env[61964]: INFO nova.compute.manager [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Terminating instance [ 851.928625] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.028508] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041122, 'name': Rename_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.204231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.204559] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.205098] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.205098] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.205313] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.207919] env[61964]: INFO nova.compute.manager [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Terminating instance [ 852.242193] env[61964]: DEBUG oslo_concurrency.lockutils [req-721da287-d51f-4610-8c0a-3d672665c961 req-84fdbe7c-f282-4060-b2b1-0030601e1e5d service nova] Releasing lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.242906] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.242906] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.346947] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.347543] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 852.350214] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.642s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.350445] env[61964]: DEBUG nova.objects.instance [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lazy-loading 'resources' on Instance uuid 2a185e4b-c636-418a-a75e-2cf158f550c0 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.404991] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 852.405824] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-551818f9-34ca-4d71-814c-5d6e4c640ac8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.421757] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 852.421757] env[61964]: value = "task-1041126" [ 852.421757] env[61964]: _type = "Task" [ 852.421757] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.435206] env[61964]: DEBUG oslo_vmware.api [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Task: {'id': task-1041124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167862} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.436136] env[61964]: DEBUG nova.compute.manager [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 852.436489] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 852.436950] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041125, 'name': PowerOffVM_Task, 'duration_secs': 0.116603} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.442668] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.443021] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 852.443361] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 852.443695] env[61964]: INFO nova.compute.manager [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Took 1.16 seconds to destroy the instance on the hypervisor. [ 852.444138] env[61964]: DEBUG oslo.service.loopingcall [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.445270] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f4b9ec-2bfb-4d6e-bad4-5fdadc3865a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.450066] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 852.450382] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 852.450784] env[61964]: DEBUG nova.compute.manager [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 852.450960] env[61964]: DEBUG nova.network.neutron [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 852.453334] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bce97c1c-6542-4af4-ae79-75dd83c61c5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.464214] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041126, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.468531] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.468971] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e86d7c97-a4b7-4bca-a78d-be86a8ec8f51 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.472165] env[61964]: DEBUG nova.network.neutron [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.481152] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 852.481152] env[61964]: value = "task-1041128" [ 852.481152] env[61964]: _type = "Task" [ 852.481152] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.490399] env[61964]: INFO nova.network.neutron [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Port fdf326b7-a14c-40cb-bfbb-569b3af8fb77 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 852.491092] env[61964]: DEBUG nova.network.neutron [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [{"id": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "address": "fa:16:3e:5d:29:ba", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0db516e7-e7", "ovs_interfaceid": "0db516e7-e7df-4355-b5d0-0d2dad89e921", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.495790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 852.496130] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 852.496410] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Deleting the datastore file [datastore2] e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.496765] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d334ca10-e504-4608-8328-edb80b084c44 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.506073] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.514868] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for the task: (returnval){ [ 852.514868] env[61964]: value = "task-1041129" [ 852.514868] env[61964]: _type = "Task" [ 852.514868] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.529929] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.533715] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041122, 'name': Rename_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.712410] env[61964]: DEBUG nova.compute.manager [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 852.712665] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 852.713570] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f7cfaf-f9fe-400b-8724-2fcf0f69a96a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.725097] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.725097] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba67fcb3-e171-4873-bab9-6ee080842c2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.731856] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 852.731856] env[61964]: value = "task-1041130" [ 852.731856] env[61964]: _type = "Task" [ 852.731856] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.740839] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041130, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.777485] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.841911] env[61964]: DEBUG nova.compute.manager [req-7f05e392-4c43-493d-ba46-84154d8e366f req-496b440f-f5a0-4219-8284-7e8c4049d730 service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-deleted-fdf326b7-a14c-40cb-bfbb-569b3af8fb77 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 852.853357] env[61964]: DEBUG nova.compute.utils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.857446] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 852.857685] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.949133] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041126, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.958503] env[61964]: DEBUG nova.policy [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1062f6da9d854397bd47603dfac77f7b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a51c67a84bd4edca0ff11769f4df0cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.960749] env[61964]: DEBUG nova.network.neutron [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [{"id": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "address": "fa:16:3e:30:78:4d", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a884ac5-73", "ovs_interfaceid": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.975459] env[61964]: DEBUG nova.network.neutron [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.996112] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-e93f98b3-4763-4f02-abd3-c24a9ab8dc17" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.998186] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041128, 'name': PowerOffVM_Task, 'duration_secs': 0.256159} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.001236] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 853.001468] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 853.003090] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a566cbd-b6b4-4fc6-ab7f-42bdde742a9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.031980] env[61964]: DEBUG oslo_vmware.api [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Task: {'id': task-1041129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122657} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.035231] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.035655] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 853.035655] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 853.035843] env[61964]: INFO nova.compute.manager [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 853.036118] env[61964]: DEBUG oslo.service.loopingcall [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.036343] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041122, 'name': Rename_Task, 'duration_secs': 1.166338} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.040175] env[61964]: DEBUG nova.compute.manager [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 853.040344] env[61964]: DEBUG nova.network.neutron [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 853.042322] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.042959] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88f4efb4-dc2b-4b0f-9800-a4b4ad3f30f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.051801] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 853.051801] env[61964]: value = "task-1041132" [ 853.051801] env[61964]: _type = "Task" [ 853.051801] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.061035] env[61964]: DEBUG nova.network.neutron [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.068022] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041132, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.106790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 853.107972] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 853.107972] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Deleting the datastore file [datastore2] eb97402e-e722-4cc3-a1d3-1b360ab1e1a7 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.107972] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fdeae1d-61ff-4a92-982d-c04957fea639 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.118821] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for the task: (returnval){ [ 853.118821] env[61964]: value = "task-1041133" [ 853.118821] env[61964]: _type = "Task" [ 853.118821] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.128093] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.208169] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d5c988-5da2-406a-a5d1-1cd20bfdec94 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.215460] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28753927-8cdb-41fa-a96e-8be279f7e19d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.253256] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1fe118-3ecc-48e7-8e4d-757fa906802f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.264164] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8f392b-79ba-423e-87a3-f8298e9a480d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.268015] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041130, 'name': PowerOffVM_Task, 'duration_secs': 0.346036} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.268327] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 853.268501] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 853.269125] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22a1dc6a-16e8-4f26-b4a4-23c4850aa6ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.280179] env[61964]: DEBUG nova.compute.provider_tree [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.306694] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Successfully created port: 3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.365827] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 853.449089] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041126, 'name': CreateSnapshot_Task, 'duration_secs': 0.843508} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.449400] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 853.450267] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce7eabd-baef-4fc3-a348-b5f93fc51924 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.465623] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.466056] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance network_info: |[{"id": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "address": "fa:16:3e:30:78:4d", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a884ac5-73", "ovs_interfaceid": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 853.466572] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:78:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a884ac5-7364-4168-bf0a-d947f84b8b92', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.481830] env[61964]: DEBUG oslo.service.loopingcall [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.482346] env[61964]: INFO nova.compute.manager [-] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Took 1.03 seconds to deallocate network for instance. [ 853.482960] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.487945] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a7fce11-5650-4429-8f22-97831e48fd24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.510720] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5713ce3d-2f82-49d8-bb03-0d6ed4dfa497 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-e93f98b3-4763-4f02-abd3-c24a9ab8dc17-15153388-0bbd-4e56-b179-42a3c5ad9552" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.791s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.517819] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.517819] env[61964]: value = "task-1041135" [ 853.517819] env[61964]: _type = "Task" [ 853.517819] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.528877] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041135, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.563845] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041132, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.566432] env[61964]: DEBUG nova.network.neutron [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.629777] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.692598] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 853.693018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 853.693316] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleting the datastore file [datastore2] e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.693649] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-413c323c-f645-4025-9747-0314b162696c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.701059] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 853.701059] env[61964]: value = "task-1041136" [ 853.701059] env[61964]: _type = "Task" [ 853.701059] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.710547] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.783173] env[61964]: DEBUG nova.scheduler.client.report [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 853.974683] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 853.975697] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ce7bfb7c-d927-4b35-9778-1891d42ea41a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.988834] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 853.988834] env[61964]: value = "task-1041137" [ 853.988834] env[61964]: _type = "Task" [ 853.988834] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.000176] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041137, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.013652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.026977] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041135, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.064562] env[61964]: DEBUG oslo_vmware.api [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041132, 'name': PowerOnVM_Task, 'duration_secs': 0.828633} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.064765] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 854.064979] env[61964]: INFO nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Took 9.53 seconds to spawn the instance on the hypervisor. [ 854.065195] env[61964]: DEBUG nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 854.065993] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6f9de6-204c-4ad4-a18d-e7ae9860cc4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.069135] env[61964]: INFO nova.compute.manager [-] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Took 1.03 seconds to deallocate network for instance. [ 854.130284] env[61964]: DEBUG oslo_vmware.api [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Task: {'id': task-1041133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.856669} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.130504] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.131142] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.131142] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.131142] env[61964]: INFO nova.compute.manager [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Took 1.69 seconds to destroy the instance on the hypervisor. [ 854.131368] env[61964]: DEBUG oslo.service.loopingcall [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.131589] env[61964]: DEBUG nova.compute.manager [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 854.131657] env[61964]: DEBUG nova.network.neutron [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 854.215236] env[61964]: DEBUG oslo_vmware.api [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041136, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.442893} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.217659] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.217870] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.218102] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.218300] env[61964]: INFO nova.compute.manager [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Took 1.51 seconds to destroy the instance on the hypervisor. [ 854.218547] env[61964]: DEBUG oslo.service.loopingcall [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.218765] env[61964]: DEBUG nova.compute.manager [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 854.218858] env[61964]: DEBUG nova.network.neutron [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 854.275847] env[61964]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port fdf326b7-a14c-40cb-bfbb-569b3af8fb77 could not be found.", "detail": ""}} {{(pid=61964) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 854.276122] env[61964]: DEBUG nova.network.neutron [-] Unable to show port fdf326b7-a14c-40cb-bfbb-569b3af8fb77 as it no longer exists. {{(pid=61964) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 854.292871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.293357] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.025s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.293600] env[61964]: DEBUG nova.objects.instance [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'resources' on Instance uuid 1c922228-4ba9-43ed-8e5c-3ca15e48ceab {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.320455] env[61964]: INFO nova.scheduler.client.report [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleted allocations for instance 2a185e4b-c636-418a-a75e-2cf158f550c0 [ 854.375965] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 854.404564] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 854.404844] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.405060] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 854.405270] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.405423] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 854.405574] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 854.405784] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 854.405972] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 854.406214] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 854.406424] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 854.406608] env[61964]: DEBUG nova.virt.hardware [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 854.407541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8f1c62-b515-441d-bbaf-3a545f8b601c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.417030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150d2271-ac67-4abc-bcb6-f82f5755c4dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.495419] env[61964]: DEBUG nova.compute.manager [req-1f720d84-4f53-492e-9258-765b52d87483 req-89a37f65-30f2-4e4d-b1e6-4cb01ec685db service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Received event network-vif-deleted-388f0e3c-60fd-49b3-b81f-5637bae949ca {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 854.495694] env[61964]: INFO nova.compute.manager [req-1f720d84-4f53-492e-9258-765b52d87483 req-89a37f65-30f2-4e4d-b1e6-4cb01ec685db service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Neutron deleted interface 388f0e3c-60fd-49b3-b81f-5637bae949ca; detaching it from the instance and deleting it from the info cache [ 854.495921] env[61964]: DEBUG nova.network.neutron [req-1f720d84-4f53-492e-9258-765b52d87483 req-89a37f65-30f2-4e4d-b1e6-4cb01ec685db service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.506370] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041137, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.528623] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041135, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.579323] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.588818] env[61964]: INFO nova.compute.manager [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Took 33.35 seconds to build instance. [ 854.830780] env[61964]: DEBUG oslo_concurrency.lockutils [None req-02e6bab7-6c4d-47c5-9ebc-266053b84a73 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "2a185e4b-c636-418a-a75e-2cf158f550c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.742s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.872911] env[61964]: DEBUG nova.compute.manager [req-d67ea6ee-8d62-42a3-ab76-371e91e3cb90 req-015566ee-34d0-4364-9a2d-4722ba5a959b service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Received event network-vif-deleted-0db516e7-e7df-4355-b5d0-0d2dad89e921 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 854.873127] env[61964]: INFO nova.compute.manager [req-d67ea6ee-8d62-42a3-ab76-371e91e3cb90 req-015566ee-34d0-4364-9a2d-4722ba5a959b service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Neutron deleted interface 0db516e7-e7df-4355-b5d0-0d2dad89e921; detaching it from the instance and deleting it from the info cache [ 854.873307] env[61964]: DEBUG nova.network.neutron [req-d67ea6ee-8d62-42a3-ab76-371e91e3cb90 req-015566ee-34d0-4364-9a2d-4722ba5a959b service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.967307] env[61964]: DEBUG nova.network.neutron [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.001152] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041137, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.001152] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef4a8d81-c655-4386-9ad1-026ef3f92105 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.012850] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039d6018-a7f7-44d2-9c8e-a56df3abdae2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.041235] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041135, 'name': CreateVM_Task, 'duration_secs': 1.351108} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.043990] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 855.045247] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.045375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.045902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.045967] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baf276e8-7fb3-4ffe-96b6-291518de58dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.056941] env[61964]: DEBUG nova.compute.manager [req-1f720d84-4f53-492e-9258-765b52d87483 req-89a37f65-30f2-4e4d-b1e6-4cb01ec685db service nova] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Detach interface failed, port_id=388f0e3c-60fd-49b3-b81f-5637bae949ca, reason: Instance eb97402e-e722-4cc3-a1d3-1b360ab1e1a7 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 855.058529] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 855.058529] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5275cb1a-b3e6-05ad-4499-569ec7b3c1c3" [ 855.058529] env[61964]: _type = "Task" [ 855.058529] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.069753] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5275cb1a-b3e6-05ad-4499-569ec7b3c1c3, 'name': SearchDatastore_Task, 'duration_secs': 0.010485} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.070088] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.072977] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.072977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.072977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.072977] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.072977] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c239f18c-c198-4ce5-a5c6-eee179a44741 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.085902] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.085902] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 855.085902] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6cd5432-85e7-4ff5-8d84-d6c408cb5b2a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.094128] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c23307fc-7022-49d4-92b7-83f7f6a5c962 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.861s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.094128] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 855.094128] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbfac5-e8ff-04c1-cd25-a8ac1bc261e7" [ 855.094128] env[61964]: _type = "Task" [ 855.094128] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.103213] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbfac5-e8ff-04c1-cd25-a8ac1bc261e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.112415] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Successfully updated port: 3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.175840] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ad98f9-0f3d-4404-bdfa-740697ca55f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.185740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af684269-2ffe-44f3-bc95-eef0ba89526d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.224347] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee5105f-217b-41a0-ac58-f4dbcec74fb9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.236725] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c32d263-3e7c-46c8-8421-f43afbd36a4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.256252] env[61964]: DEBUG nova.compute.provider_tree [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.319445] env[61964]: DEBUG nova.network.neutron [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.376272] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2950a0c-d3cb-4799-b654-4296a8aa28fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.386720] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101c8406-c341-4da5-80dd-f5faa4c658cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.418130] env[61964]: DEBUG nova.compute.manager [req-d67ea6ee-8d62-42a3-ab76-371e91e3cb90 req-015566ee-34d0-4364-9a2d-4722ba5a959b service nova] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Detach interface failed, port_id=0db516e7-e7df-4355-b5d0-0d2dad89e921, reason: Instance e93f98b3-4763-4f02-abd3-c24a9ab8dc17 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 855.474765] env[61964]: INFO nova.compute.manager [-] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Took 1.34 seconds to deallocate network for instance. [ 855.499759] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041137, 'name': CloneVM_Task, 'duration_secs': 1.441506} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.500117] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Created linked-clone VM from snapshot [ 855.500879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbce5b09-482f-4444-ad60-7a890a855868 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.509683] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Uploading image f732a7b4-4004-478c-ad84-b0f50659c288 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 855.537293] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 855.537293] env[61964]: value = "vm-230504" [ 855.537293] env[61964]: _type = "VirtualMachine" [ 855.537293] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 855.537583] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-810429ff-71b0-4e9e-a82b-1b992d08bd86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.546341] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lease: (returnval){ [ 855.546341] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526838b0-3622-6a92-619f-e61f5e1ad4b1" [ 855.546341] env[61964]: _type = "HttpNfcLease" [ 855.546341] env[61964]: } obtained for exporting VM: (result){ [ 855.546341] env[61964]: value = "vm-230504" [ 855.546341] env[61964]: _type = "VirtualMachine" [ 855.546341] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 855.546341] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the lease: (returnval){ [ 855.546341] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526838b0-3622-6a92-619f-e61f5e1ad4b1" [ 855.546341] env[61964]: _type = "HttpNfcLease" [ 855.546341] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 855.558211] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 855.558211] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526838b0-3622-6a92-619f-e61f5e1ad4b1" [ 855.558211] env[61964]: _type = "HttpNfcLease" [ 855.558211] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 855.558512] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 855.558512] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526838b0-3622-6a92-619f-e61f5e1ad4b1" [ 855.558512] env[61964]: _type = "HttpNfcLease" [ 855.558512] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 855.559255] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982d8241-358b-486e-831b-fbbbed84d7aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.567183] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 855.567387] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 855.626998] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.627169] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquired lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.627325] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.646127] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dbfac5-e8ff-04c1-cd25-a8ac1bc261e7, 'name': SearchDatastore_Task, 'duration_secs': 0.011231} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.647429] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1a0030d-d052-4f0b-ac87-c40a348bb591 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.654083] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 855.654083] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbb2d-4f4d-5314-23a6-f1ebe84b699d" [ 855.654083] env[61964]: _type = "Task" [ 855.654083] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.663165] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbb2d-4f4d-5314-23a6-f1ebe84b699d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.672573] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ffe13d14-e402-44d3-94d3-f8e5c28c8728 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.682579] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.759607] env[61964]: DEBUG nova.scheduler.client.report [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.822499] env[61964]: INFO nova.compute.manager [-] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Took 1.60 seconds to deallocate network for instance. [ 855.921733] env[61964]: DEBUG nova.network.neutron [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Updating instance_info_cache with network_info: [{"id": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "address": "fa:16:3e:fb:62:72", "network": {"id": "b65bdb6d-1434-4605-8db9-17052ebacf32", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-960847416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a51c67a84bd4edca0ff11769f4df0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4f7d05-0a", "ovs_interfaceid": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.981747] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.139834] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.140263] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.141240] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.141240] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.141409] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.143516] env[61964]: INFO nova.compute.manager [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Terminating instance [ 856.167317] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbb2d-4f4d-5314-23a6-f1ebe84b699d, 'name': SearchDatastore_Task, 'duration_secs': 0.010475} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.167367] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.167711] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 856.168114] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f73b906-b9c9-4d4e-b1f3-8a5e32964c2b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.176871] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 856.176871] env[61964]: value = "task-1041139" [ 856.176871] env[61964]: _type = "Task" [ 856.176871] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.186699] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041139, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.266049] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.269413] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.190s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.269413] env[61964]: DEBUG nova.objects.instance [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lazy-loading 'resources' on Instance uuid b90e7ebd-8dc6-495b-99bf-15ff24763af4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.292351] env[61964]: INFO nova.scheduler.client.report [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance 1c922228-4ba9-43ed-8e5c-3ca15e48ceab [ 856.330801] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.425465] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Releasing lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.425803] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Instance network_info: |[{"id": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "address": "fa:16:3e:fb:62:72", "network": {"id": "b65bdb6d-1434-4605-8db9-17052ebacf32", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-960847416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a51c67a84bd4edca0ff11769f4df0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4f7d05-0a", "ovs_interfaceid": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 856.426409] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:62:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '352165bb-004f-4180-9627-3a275dbe18af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e4f7d05-0a63-48d8-9a6f-6283822f3f9a', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.435082] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Creating folder: Project (5a51c67a84bd4edca0ff11769f4df0cb). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 856.436727] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aea9f822-e16d-4d10-900e-36df6ff5362d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.453928] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Created folder: Project (5a51c67a84bd4edca0ff11769f4df0cb) in parent group-v230360. [ 856.454322] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Creating folder: Instances. Parent ref: group-v230505. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 856.454614] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41f721dc-3b57-459a-80fd-057a95aba404 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.470919] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Created folder: Instances in parent group-v230505. [ 856.471389] env[61964]: DEBUG oslo.service.loopingcall [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.471740] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.472069] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d0312d1-da6c-43d3-8b79-5617eabb07b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.498142] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.498142] env[61964]: value = "task-1041142" [ 856.498142] env[61964]: _type = "Task" [ 856.498142] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.508659] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041142, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.656721] env[61964]: DEBUG nova.compute.manager [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 856.657266] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 856.658765] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0900a573-7f63-4d7f-bf39-73b1e59ee5b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.672906] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.674215] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10e94f3f-dcdc-41fa-8474-c40a6330b211 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.685031] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 856.685031] env[61964]: value = "task-1041143" [ 856.685031] env[61964]: _type = "Task" [ 856.685031] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.698291] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041139, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50811} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.699474] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.699948] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.700444] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79133220-d943-45c2-b4f2-5beeb4cbda36 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.708992] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.715313] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 856.715313] env[61964]: value = "task-1041144" [ 856.715313] env[61964]: _type = "Task" [ 856.715313] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.728237] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.800038] env[61964]: DEBUG oslo_concurrency.lockutils [None req-561ddd2c-c0f7-4568-b9bb-3ea066f4a88b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "1c922228-4ba9-43ed-8e5c-3ca15e48ceab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.429s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.922212] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.922663] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.922948] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.923229] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.923923] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.926461] env[61964]: INFO nova.compute.manager [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Terminating instance [ 856.957633] env[61964]: DEBUG nova.compute.manager [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Received event network-vif-plugged-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 856.958338] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Acquiring lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.958338] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.958451] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.958621] env[61964]: DEBUG nova.compute.manager [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] No waiting events found dispatching network-vif-plugged-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 856.958789] env[61964]: WARNING nova.compute.manager [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Received unexpected event network-vif-plugged-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a for instance with vm_state building and task_state spawning. [ 856.958973] env[61964]: DEBUG nova.compute.manager [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Received event network-changed-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 856.959162] env[61964]: DEBUG nova.compute.manager [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Refreshing instance network info cache due to event network-changed-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 856.959357] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Acquiring lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.959570] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Acquired lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.959746] env[61964]: DEBUG nova.network.neutron [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Refreshing network info cache for port 3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 857.010440] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041142, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.062670] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ca59c4-06a1-404c-843c-09d66b5381fb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.071808] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232f8472-6108-4c9d-8372-1d643d08daf2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.111798] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba80d28-851a-45af-b3fd-495780ccb3e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.121626] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68899b6a-9d8c-41f2-be5b-f37cad1eb6b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.140840] env[61964]: DEBUG nova.compute.provider_tree [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.196858] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041143, 'name': PowerOffVM_Task, 'duration_secs': 0.301085} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.197357] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 857.198904] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 857.199364] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9753df1b-f5a4-45bf-b980-51071d536807 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.228067] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111332} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.228560] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.229569] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4fd671-51d0-40b1-8a38-992d27c8e66b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.254403] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.255027] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dca639c-7f7f-43b8-bbd9-42f4c6b776e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.283321] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 857.283321] env[61964]: value = "task-1041146" [ 857.283321] env[61964]: _type = "Task" [ 857.283321] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.289173] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 857.289719] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 857.290099] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore2] fb96bf3a-ea2c-414d-bb29-eca2b07d51b6 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.290891] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf7feaf5-bd39-4557-99b3-a404f7e350e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.297537] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041146, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.303211] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 857.303211] env[61964]: value = "task-1041147" [ 857.303211] env[61964]: _type = "Task" [ 857.303211] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.313754] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.430604] env[61964]: DEBUG nova.compute.manager [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 857.430922] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 857.432093] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc27ab5-8bc0-4d4b-bf4b-32e87a571b1c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.441250] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 857.441701] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d44fd71-bd91-467d-9c92-ed54a2cf5164 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.449700] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 857.449700] env[61964]: value = "task-1041148" [ 857.449700] env[61964]: _type = "Task" [ 857.449700] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.461915] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.512884] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041142, 'name': CreateVM_Task, 'duration_secs': 0.532523} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.512884] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 857.513757] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.514095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.514710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.516015] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ee5adc9-3a8b-4c70-8858-a7db1be562b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.522126] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 857.522126] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5279286a-71da-e5cd-f9d8-59661c5c3590" [ 857.522126] env[61964]: _type = "Task" [ 857.522126] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.532955] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5279286a-71da-e5cd-f9d8-59661c5c3590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.644109] env[61964]: DEBUG nova.scheduler.client.report [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.798705] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041146, 'name': ReconfigVM_Task, 'duration_secs': 0.437972} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.800546] env[61964]: DEBUG nova.network.neutron [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Updated VIF entry in instance network info cache for port 3e4f7d05-0a63-48d8-9a6f-6283822f3f9a. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.801423] env[61964]: DEBUG nova.network.neutron [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Updating instance_info_cache with network_info: [{"id": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "address": "fa:16:3e:fb:62:72", "network": {"id": "b65bdb6d-1434-4605-8db9-17052ebacf32", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-960847416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a51c67a84bd4edca0ff11769f4df0cb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e4f7d05-0a", "ovs_interfaceid": "3e4f7d05-0a63-48d8-9a6f-6283822f3f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.803385] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.807130] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3ae2d0f-fd0b-44b1-912a-a0d70c99bc28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.820605] env[61964]: DEBUG oslo_vmware.api [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194606} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.822819] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.823514] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.823790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.824381] env[61964]: INFO nova.compute.manager [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 857.824381] env[61964]: DEBUG oslo.service.loopingcall [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.825037] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 857.825037] env[61964]: value = "task-1041149" [ 857.825037] env[61964]: _type = "Task" [ 857.825037] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.825319] env[61964]: DEBUG nova.compute.manager [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 857.826652] env[61964]: DEBUG nova.network.neutron [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 857.838412] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041149, 'name': Rename_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.965836] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041148, 'name': PowerOffVM_Task, 'duration_secs': 0.234872} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.966284] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 857.966532] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 857.967105] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74253906-f94e-4a0e-943d-88b7107d12cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.034591] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5279286a-71da-e5cd-f9d8-59661c5c3590, 'name': SearchDatastore_Task, 'duration_secs': 0.012827} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.034910] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.035227] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.035461] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.037337] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.037337] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.037337] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfa82130-d6c6-4f3f-b40f-e4ba8c36c74f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.043198] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 858.043465] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 858.043659] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleting the datastore file [datastore1] 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.044048] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e1a88a5-5fc6-4dc2-b27f-8e345d95dcd3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.049159] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.049365] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 858.050223] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce43e4fb-a4d0-478e-ac71-89336c4cf7de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.054275] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for the task: (returnval){ [ 858.054275] env[61964]: value = "task-1041151" [ 858.054275] env[61964]: _type = "Task" [ 858.054275] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.061634] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 858.061634] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52045fe2-493c-26d9-14d5-324555fd8e09" [ 858.061634] env[61964]: _type = "Task" [ 858.061634] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.068614] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.075027] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52045fe2-493c-26d9-14d5-324555fd8e09, 'name': SearchDatastore_Task, 'duration_secs': 0.010893} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.075379] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70f64d3e-337b-4e4c-849b-c233e8b131ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.083464] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 858.083464] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52406cee-d3f3-3d55-4e7a-60c819850fe7" [ 858.083464] env[61964]: _type = "Task" [ 858.083464] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.094815] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52406cee-d3f3-3d55-4e7a-60c819850fe7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.151055] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.155518] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.598s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.155854] env[61964]: DEBUG nova.objects.instance [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lazy-loading 'resources' on Instance uuid 16276be9-b305-4d1d-afde-bc98be42687a {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.173733] env[61964]: INFO nova.scheduler.client.report [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleted allocations for instance b90e7ebd-8dc6-495b-99bf-15ff24763af4 [ 858.306968] env[61964]: DEBUG oslo_concurrency.lockutils [req-d597c6ab-aa49-4288-8de3-4a3dcd1e2bab req-3056cbd0-14b6-4e1d-911e-c20520f6739a service nova] Releasing lock "refresh_cache-a1e5c369-4233-4794-adda-1ba1e89e8154" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.338893] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041149, 'name': Rename_Task, 'duration_secs': 0.149137} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.339218] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 858.339481] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d740693-c7b6-4545-b8b9-c312f142bb18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.347864] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 858.347864] env[61964]: value = "task-1041152" [ 858.347864] env[61964]: _type = "Task" [ 858.347864] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.356738] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.567449] env[61964]: DEBUG oslo_vmware.api [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Task: {'id': task-1041151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140712} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.567697] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.567942] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 858.568216] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 858.568408] env[61964]: INFO nova.compute.manager [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 858.568660] env[61964]: DEBUG oslo.service.loopingcall [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.568860] env[61964]: DEBUG nova.compute.manager [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 858.568955] env[61964]: DEBUG nova.network.neutron [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 858.592924] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52406cee-d3f3-3d55-4e7a-60c819850fe7, 'name': SearchDatastore_Task, 'duration_secs': 0.011139} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.593235] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.593498] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a1e5c369-4233-4794-adda-1ba1e89e8154/a1e5c369-4233-4794-adda-1ba1e89e8154.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.593760] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1573e69-e6ff-4642-9b89-0a648e72a21c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.602127] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 858.602127] env[61964]: value = "task-1041153" [ 858.602127] env[61964]: _type = "Task" [ 858.602127] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.606896] env[61964]: DEBUG nova.network.neutron [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.612020] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.685519] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9d6638fc-5cac-4f37-8d46-ca84813d17f4 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "b90e7ebd-8dc6-495b-99bf-15ff24763af4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.408s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.860903] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.937792] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045abab4-e6ae-43cb-ac7e-943e2933a678 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.950424] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a281466b-1acd-43f2-9365-6468d3bb444e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.986092] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2140a5d3-50ff-4612-97bc-ccdd1af9eec0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.996547] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f2381d-e9f9-418e-8c71-a3184d8a887c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.015302] env[61964]: DEBUG nova.compute.provider_tree [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.113567] env[61964]: INFO nova.compute.manager [-] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Took 1.29 seconds to deallocate network for instance. [ 859.113567] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475009} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.118853] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a1e5c369-4233-4794-adda-1ba1e89e8154/a1e5c369-4233-4794-adda-1ba1e89e8154.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.119150] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.124841] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57064c67-be5e-49ac-92c3-797300508a5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.137565] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 859.137565] env[61964]: value = "task-1041154" [ 859.137565] env[61964]: _type = "Task" [ 859.137565] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.147171] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.190949] env[61964]: DEBUG nova.compute.manager [req-68ec501c-1be0-4c07-9051-5c3fead061e7 req-f174cbc1-349c-4a78-9f7a-ccbc1569f319 service nova] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Received event network-vif-deleted-6467ee6e-8b51-47b3-8441-d53dd716c656 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 859.360094] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041152, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.535812] env[61964]: ERROR nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] [req-a78258be-66db-49e3-9eef-2f15c8355b48] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a78258be-66db-49e3-9eef-2f15c8355b48"}]} [ 859.551104] env[61964]: DEBUG nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 859.565277] env[61964]: DEBUG nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 859.565529] env[61964]: DEBUG nova.compute.provider_tree [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.577203] env[61964]: DEBUG nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 859.594496] env[61964]: DEBUG nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 859.629362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.635097] env[61964]: DEBUG nova.network.neutron [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.650008] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088025} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.650890] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.651823] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e00cd3-a741-42db-bb0c-87cb2edcee47 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.678539] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] a1e5c369-4233-4794-adda-1ba1e89e8154/a1e5c369-4233-4794-adda-1ba1e89e8154.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.681731] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-641e2fc7-d0b8-46e7-a3e2-502bf76b326d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.700935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.701237] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.704032] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 859.704032] env[61964]: value = "task-1041156" [ 859.704032] env[61964]: _type = "Task" [ 859.704032] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.717341] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.819424] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.820012] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.864518] env[61964]: DEBUG oslo_vmware.api [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041152, 'name': PowerOnVM_Task, 'duration_secs': 1.141943} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.864815] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 859.865031] env[61964]: INFO nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Took 9.93 seconds to spawn the instance on the hypervisor. [ 859.865226] env[61964]: DEBUG nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 859.866017] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9daf5d7-a56f-4c22-93a3-1108a1a8055c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.881602] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167fe73b-8e10-4ec2-a747-3c3186a205f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.890501] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f98d2c-a5df-4c4e-9809-39c612416826 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.929302] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9aca7c6-be49-4b5f-b381-a6f64b871123 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.938259] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3861f7-23a9-4564-a13d-cd4988ae06cc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.954503] env[61964]: DEBUG nova.compute.provider_tree [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.137692] env[61964]: INFO nova.compute.manager [-] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Took 1.57 seconds to deallocate network for instance. [ 860.205796] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 860.218352] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041156, 'name': ReconfigVM_Task, 'duration_secs': 0.319035} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.218740] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Reconfigured VM instance instance-0000004b to attach disk [datastore1] a1e5c369-4233-4794-adda-1ba1e89e8154/a1e5c369-4233-4794-adda-1ba1e89e8154.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.219748] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dfbf7377-c518-4fec-8d5c-a52670de1225 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.228934] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 860.228934] env[61964]: value = "task-1041157" [ 860.228934] env[61964]: _type = "Task" [ 860.228934] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.243858] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041157, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.327925] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 860.327925] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 860.386319] env[61964]: INFO nova.compute.manager [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Took 30.04 seconds to build instance. [ 860.486648] env[61964]: DEBUG nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 94 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 860.488150] env[61964]: DEBUG nova.compute.provider_tree [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 94 to 95 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 860.488150] env[61964]: DEBUG nova.compute.provider_tree [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.645097] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.731147] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.742153] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041157, 'name': Rename_Task, 'duration_secs': 0.145635} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.742444] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.742713] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-703ca784-8bf1-4b08-b125-a1b27a6a230c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.750247] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 860.750247] env[61964]: value = "task-1041158" [ 860.750247] env[61964]: _type = "Task" [ 860.750247] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.759060] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.888331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6cb9a18f-f76b-4b7a-83ea-9f8a7176175e tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.550s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.904721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.904875] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.905057] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 860.993052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.839s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.995519] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.142s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.997789] env[61964]: INFO nova.compute.claims [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.013874] env[61964]: INFO nova.scheduler.client.report [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Deleted allocations for instance 16276be9-b305-4d1d-afde-bc98be42687a [ 861.222135] env[61964]: DEBUG nova.compute.manager [req-86e116cb-b0cf-4102-beb9-7302899f0acc req-d10c7073-d8bf-4be6-8b0f-6b32cd06639b service nova] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Received event network-vif-deleted-1e892208-7628-4004-8011-abf5a4a211fc {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 861.261056] env[61964]: DEBUG oslo_vmware.api [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041158, 'name': PowerOnVM_Task, 'duration_secs': 0.472103} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.261349] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.261575] env[61964]: INFO nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Took 6.89 seconds to spawn the instance on the hypervisor. [ 861.261752] env[61964]: DEBUG nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.262611] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea2d6ea-a4dc-46e0-8f39-1172a81994cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.522778] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bfc5f37d-823c-46ac-bcf6-1b317a4d1857 tempest-ServerRescueTestJSON-1075548192 tempest-ServerRescueTestJSON-1075548192-project-member] Lock "16276be9-b305-4d1d-afde-bc98be42687a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.399s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.783553] env[61964]: INFO nova.compute.manager [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Took 29.62 seconds to build instance. [ 861.934551] env[61964]: DEBUG nova.compute.manager [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Stashing vm_state: active {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 862.288220] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0892a9b1-6c92-4c0a-8136-df7619d3efb1 tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.138s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.313280] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d787ddc6-d21c-472a-92be-52fffb8ee0df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.322149] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ca17f4-0b2d-4713-8415-3188502ea1a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.357112] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee65205-2c0a-451c-a9d5-58a9c072f423 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.366612] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c74491-87ef-4931-b703-bc1edf027bc4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.382137] env[61964]: DEBUG nova.compute.provider_tree [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.430941] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "a1e5c369-4233-4794-adda-1ba1e89e8154" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.431271] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.431550] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.431759] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.431940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.434335] env[61964]: INFO nova.compute.manager [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Terminating instance [ 862.440191] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating instance_info_cache with network_info: [{"id": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "address": "fa:16:3e:64:d3:d6", "network": {"id": "e56cb021-8c44-423a-94c1-804c7c88a7d0", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-804484474-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f864c21d43647f194c05c358e2dc66c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bab6a6c3-1c5c-4776-b21b-dec21196d702", "external-id": "nsx-vlan-transportzone-634", "segmentation_id": 634, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap822d4088-78", "ovs_interfaceid": "822d4088-7886-4e3d-8716-59e4a3a18ab0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.454941] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.888755] env[61964]: DEBUG nova.scheduler.client.report [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 862.938225] env[61964]: DEBUG nova.compute.manager [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 862.938656] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.940181] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b5360e-581d-42bd-a220-e6cce158cb74 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.943939] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.944288] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 862.944529] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.945201] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.945378] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.945537] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.945713] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.945864] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.946064] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 862.946226] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 862.950614] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.951212] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8bebf75f-3703-4fe4-9dc8-1973cd123d11 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.960508] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 862.960508] env[61964]: value = "task-1041159" [ 862.960508] env[61964]: _type = "Task" [ 862.960508] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.970642] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.042100] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.042434] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.042914] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.043273] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.043359] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.045767] env[61964]: INFO nova.compute.manager [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Terminating instance [ 863.394403] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.395035] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 863.397676] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.053s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.449386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.471772] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041159, 'name': PowerOffVM_Task, 'duration_secs': 0.197976} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.472550] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.472815] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.473099] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a4de707-2801-4596-bf20-adde18d8310c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.545653] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.545885] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.546136] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Deleting the datastore file [datastore1] a1e5c369-4233-4794-adda-1ba1e89e8154 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.546464] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-936d3570-7246-451e-89d6-fbe9638320c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.550029] env[61964]: DEBUG nova.compute.manager [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 863.550269] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 863.551234] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e01f3d-a8a9-4c19-a435-d3338f34b7c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.555323] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for the task: (returnval){ [ 863.555323] env[61964]: value = "task-1041161" [ 863.555323] env[61964]: _type = "Task" [ 863.555323] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.561248] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 863.561855] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db05e3ab-c546-4c2e-8532-4ca0bc4d90d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.567156] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.572342] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 863.572342] env[61964]: value = "task-1041162" [ 863.572342] env[61964]: _type = "Task" [ 863.572342] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.581539] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.902802] env[61964]: DEBUG nova.compute.utils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.905912] env[61964]: INFO nova.compute.claims [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.911125] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 863.911125] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 863.961412] env[61964]: DEBUG nova.policy [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05bfff4e5dae488593812635d1c216b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f1804a0ded74317a1e2016db18c55ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.069546] env[61964]: DEBUG oslo_vmware.api [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Task: {'id': task-1041161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196696} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.069995] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.070192] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.070810] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.071610] env[61964]: INFO nova.compute.manager [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Took 1.13 seconds to destroy the instance on the hypervisor. [ 864.072037] env[61964]: DEBUG oslo.service.loopingcall [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.072386] env[61964]: DEBUG nova.compute.manager [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 864.072878] env[61964]: DEBUG nova.network.neutron [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.086820] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041162, 'name': PowerOffVM_Task, 'duration_secs': 0.273151} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.087150] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.087341] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 864.088329] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e4e9348-00e6-4b2c-8a92-1b62aa651f02 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.165586] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 864.165891] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 864.166156] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleting the datastore file [datastore2] a995ff8c-75b9-4089-ad4c-9e6baa33fe8c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.166460] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d09dd560-c012-4ab0-bb43-38103229fa85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.175171] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for the task: (returnval){ [ 864.175171] env[61964]: value = "task-1041164" [ 864.175171] env[61964]: _type = "Task" [ 864.175171] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.188850] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.268314] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 864.269511] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a874e1-5f61-4ecf-8f79-7d6426aae083 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.277096] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 864.277096] env[61964]: ERROR oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk due to incomplete transfer. [ 864.278539] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9d77c712-d1ab-4ed6-8194-d93e7755984a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.285552] env[61964]: DEBUG oslo_vmware.rw_handles [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524ec5f8-63f4-42a1-6157-8a491fb0d9fa/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 864.285765] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Uploaded image f732a7b4-4004-478c-ad84-b0f50659c288 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 864.289154] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 864.289574] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-929d1fcb-bf0a-432f-8422-7c7d747e1ab0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.298062] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 864.298062] env[61964]: value = "task-1041165" [ 864.298062] env[61964]: _type = "Task" [ 864.298062] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.311911] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041165, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.403179] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Successfully created port: 31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.411697] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 864.422624] env[61964]: INFO nova.compute.resource_tracker [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating resource usage from migration 8eb2d8ea-8ce7-4c30-9e2e-6127519d71b3 [ 864.512090] env[61964]: DEBUG nova.compute.manager [req-eec831db-93c2-4e0e-a8b1-194755736bbc req-b0c07738-4973-4158-b318-07c4407fa081 service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Received event network-vif-deleted-3e4f7d05-0a63-48d8-9a6f-6283822f3f9a {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 864.512348] env[61964]: INFO nova.compute.manager [req-eec831db-93c2-4e0e-a8b1-194755736bbc req-b0c07738-4973-4158-b318-07c4407fa081 service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Neutron deleted interface 3e4f7d05-0a63-48d8-9a6f-6283822f3f9a; detaching it from the instance and deleting it from the info cache [ 864.512534] env[61964]: DEBUG nova.network.neutron [req-eec831db-93c2-4e0e-a8b1-194755736bbc req-b0c07738-4973-4158-b318-07c4407fa081 service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.686591] env[61964]: DEBUG oslo_vmware.api [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Task: {'id': task-1041164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155349} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.689847] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.690347] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 864.690922] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.691355] env[61964]: INFO nova.compute.manager [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 864.691752] env[61964]: DEBUG oslo.service.loopingcall [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.692432] env[61964]: DEBUG nova.compute.manager [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 864.692642] env[61964]: DEBUG nova.network.neutron [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.718496] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8af94b9-239a-4f3c-be2a-5e8cdc9038fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.728657] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfbcef9-da44-43f4-9787-a059d70d2a1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.770864] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447a5308-bb76-4578-9063-947df5a54091 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.780659] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88917ffd-68f8-4ed9-afdd-cda3374745c2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.796499] env[61964]: DEBUG nova.compute.provider_tree [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.809333] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041165, 'name': Destroy_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.975837] env[61964]: DEBUG nova.network.neutron [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.015437] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23cdcaae-ada3-441a-95a9-a1432ab33e55 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.029365] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655181ad-848e-47e7-a349-31b6b1382af1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.062307] env[61964]: DEBUG nova.compute.manager [req-eec831db-93c2-4e0e-a8b1-194755736bbc req-b0c07738-4973-4158-b318-07c4407fa081 service nova] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Detach interface failed, port_id=3e4f7d05-0a63-48d8-9a6f-6283822f3f9a, reason: Instance a1e5c369-4233-4794-adda-1ba1e89e8154 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 865.080602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.080862] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.301201] env[61964]: DEBUG nova.scheduler.client.report [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 865.314365] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041165, 'name': Destroy_Task, 'duration_secs': 0.890495} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.314637] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Destroyed the VM [ 865.314917] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 865.315745] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0afdd075-ee57-491c-ab51-fad765cc4234 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.322041] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "ba0f227f-640a-4765-a84e-61bd81128356" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.322307] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.324693] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 865.324693] env[61964]: value = "task-1041166" [ 865.324693] env[61964]: _type = "Task" [ 865.324693] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.335727] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.425801] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 865.453281] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b2ac7b2beea6c491cb124ea94cd90d97',container_format='bare',created_at=2024-09-30T16:25:14Z,direct_url=,disk_format='vmdk',id=acef627a-6665-4737-891e-74c2350a5188,min_disk=1,min_ram=0,name='tempest-test-snap-321621508',owner='3f1804a0ded74317a1e2016db18c55ca',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-30T16:25:31Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 865.453575] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.453743] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 865.453963] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.454164] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 865.454362] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 865.454586] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 865.454758] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 865.454929] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 865.455119] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 865.455304] env[61964]: DEBUG nova.virt.hardware [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 865.456515] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b214943-d7b2-4e82-ac4a-494c08113076 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.466416] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401f0a2f-d214-46c5-bd2a-7977e1ab7be2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.485617] env[61964]: INFO nova.compute.manager [-] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Took 1.41 seconds to deallocate network for instance. [ 865.583059] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 865.808940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.411s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.809183] env[61964]: INFO nova.compute.manager [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Migrating [ 865.816611] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.746s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.816816] env[61964]: DEBUG nova.objects.instance [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 865.828014] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 865.839546] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.847363] env[61964]: DEBUG nova.network.neutron [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.994221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.110343] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.198174] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Successfully updated port: 31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.331240] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.331501] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.331707] env[61964]: DEBUG nova.network.neutron [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.345883] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041166, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.348705] env[61964]: INFO nova.compute.manager [-] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Took 1.66 seconds to deallocate network for instance. [ 866.355872] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.540618] env[61964]: DEBUG nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Received event network-vif-deleted-822d4088-7886-4e3d-8716-59e4a3a18ab0 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 866.540899] env[61964]: DEBUG nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Received event network-vif-plugged-31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 866.541127] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Acquiring lock "50b04cbf-9064-4010-b350-638e8096bb96-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.541340] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Lock "50b04cbf-9064-4010-b350-638e8096bb96-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.541519] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Lock "50b04cbf-9064-4010-b350-638e8096bb96-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.541639] env[61964]: DEBUG nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] No waiting events found dispatching network-vif-plugged-31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 866.541812] env[61964]: WARNING nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Received unexpected event network-vif-plugged-31b39082-5123-4292-bb5a-18184dc24861 for instance with vm_state building and task_state spawning. [ 866.542042] env[61964]: DEBUG nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Received event network-changed-31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 866.542192] env[61964]: DEBUG nova.compute.manager [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Refreshing instance network info cache due to event network-changed-31b39082-5123-4292-bb5a-18184dc24861. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 866.542405] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Acquiring lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.542548] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Acquired lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.542730] env[61964]: DEBUG nova.network.neutron [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Refreshing network info cache for port 31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.700913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.835509] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5fb0132e-d572-4ebf-9149-9c6252766f65 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.836592] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.676s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.836791] env[61964]: DEBUG nova.objects.instance [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 866.847039] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041166, 'name': RemoveSnapshot_Task, 'duration_secs': 1.280899} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.847322] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 866.848648] env[61964]: DEBUG nova.compute.manager [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 866.849235] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cfa9df-9d81-4c3c-a385-68551381ce1c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.854172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.318944] env[61964]: DEBUG nova.network.neutron [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.329845] env[61964]: DEBUG nova.network.neutron [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.365872] env[61964]: INFO nova.compute.manager [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Shelve offloading [ 867.455380] env[61964]: DEBUG nova.network.neutron [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.836892] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.849128] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c5061e61-d2ad-4a84-8ff6-8a6c1b34ad33 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.850268] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.868s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.850489] env[61964]: DEBUG nova.objects.instance [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lazy-loading 'resources' on Instance uuid 46435c68-f85a-4360-b2b5-6296afc33c3e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.869358] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.869644] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7c04a22-e8ac-4155-9bdd-6c80084c4eda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.879187] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 867.879187] env[61964]: value = "task-1041167" [ 867.879187] env[61964]: _type = "Task" [ 867.879187] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.887783] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.957708] env[61964]: DEBUG oslo_concurrency.lockutils [req-4fe86d90-0555-41f6-8502-8b50a2b028f5 req-d182716d-b22f-43d5-8635-50ff8e2a41ac service nova] Releasing lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.958208] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.958424] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.393845] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 868.393845] env[61964]: DEBUG nova.compute.manager [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 868.393845] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad96263-3c4a-4977-b455-940dfcb75242 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.399445] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.399721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.399947] env[61964]: DEBUG nova.network.neutron [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.493374] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.637915] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32758224-5c03-4e4e-affb-1cb85b4f79ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.646162] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fdca0e-895e-4446-ac92-a7f3925a646c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.679887] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3f9a02-6965-4524-a28c-cbb3b294726c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.685092] env[61964]: DEBUG nova.network.neutron [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Updating instance_info_cache with network_info: [{"id": "31b39082-5123-4292-bb5a-18184dc24861", "address": "fa:16:3e:01:18:36", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31b39082-51", "ovs_interfaceid": "31b39082-5123-4292-bb5a-18184dc24861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.690336] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36993cc8-11fd-40e4-907a-2c479686d4e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.704458] env[61964]: DEBUG nova.compute.provider_tree [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.093744] env[61964]: DEBUG nova.network.neutron [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.186387] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "refresh_cache-50b04cbf-9064-4010-b350-638e8096bb96" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.186713] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance network_info: |[{"id": "31b39082-5123-4292-bb5a-18184dc24861", "address": "fa:16:3e:01:18:36", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31b39082-51", "ovs_interfaceid": "31b39082-5123-4292-bb5a-18184dc24861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 869.187176] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:18:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31b39082-5123-4292-bb5a-18184dc24861', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.194819] env[61964]: DEBUG oslo.service.loopingcall [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.195055] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 869.195283] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c380ea47-688e-4f16-a957-1d2aeb8f4ab0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.210560] env[61964]: DEBUG nova.scheduler.client.report [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.219668] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.219668] env[61964]: value = "task-1041168" [ 869.219668] env[61964]: _type = "Task" [ 869.219668] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.229534] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041168, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.353827] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd61e16-e0a0-42e0-a534-eb72ab7a3450 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.373438] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 869.548013] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.548302] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.598047] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.715143] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.717646] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.704s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.717745] env[61964]: DEBUG nova.objects.instance [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lazy-loading 'resources' on Instance uuid 93720ce5-834a-4d8d-b230-13d38620b688 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.730628] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041168, 'name': CreateVM_Task, 'duration_secs': 0.301763} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.730768] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 869.731415] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.732119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.732119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 869.732295] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f93f0de-4c03-47e2-a5c8-f787f23a2faa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.737798] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 869.737798] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52bf83c4-c777-c299-d92b-277af8a352d6" [ 869.737798] env[61964]: _type = "Task" [ 869.737798] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.738772] env[61964]: INFO nova.scheduler.client.report [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Deleted allocations for instance 46435c68-f85a-4360-b2b5-6296afc33c3e [ 869.756161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.756161] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Processing image acef627a-6665-4737-891e-74c2350a5188 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.756161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.756161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.756437] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.756714] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-823e75b9-eff5-4957-9289-a97641dd5c48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.767549] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.767741] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 869.768530] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd1f175a-dc08-4dc3-a145-82a6d305a77d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.774833] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 869.774833] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c45bfc-4c7e-db45-3305-05f37d32949f" [ 869.774833] env[61964]: _type = "Task" [ 869.774833] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.784223] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c45bfc-4c7e-db45-3305-05f37d32949f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.835431] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 869.836348] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f79730-fbcc-4643-9422-66c8bf8321d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.844658] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.844915] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6dca2ed-342d-4bcb-b144-11e1b14197a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.880026] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 869.880533] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfca46ed-a068-4d1f-920a-a85ad1499a58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.888422] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 869.888422] env[61964]: value = "task-1041170" [ 869.888422] env[61964]: _type = "Task" [ 869.888422] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.898495] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 869.898818] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 869.913677] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.913851] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.913932] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleting the datastore file [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.914226] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d701c0a-4139-4093-ba5c-9e8b0ca20b13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.922193] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 869.922193] env[61964]: value = "task-1041171" [ 869.922193] env[61964]: _type = "Task" [ 869.922193] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.933205] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041171, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.994733] env[61964]: DEBUG nova.compute.manager [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-vif-unplugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 869.995015] env[61964]: DEBUG oslo_concurrency.lockutils [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.995292] env[61964]: DEBUG oslo_concurrency.lockutils [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.995519] env[61964]: DEBUG oslo_concurrency.lockutils [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.995752] env[61964]: DEBUG nova.compute.manager [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] No waiting events found dispatching network-vif-unplugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 869.995957] env[61964]: WARNING nova.compute.manager [req-21807a3f-b224-427a-9a62-9de6548f8168 req-42b34e70-3d10-4539-a26e-7058892b4a81 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received unexpected event network-vif-unplugged-edda620f-1259-4fbb-afa6-48aef4eda40b for instance with vm_state shelved and task_state shelving_offloading. [ 870.051861] env[61964]: DEBUG nova.compute.utils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.252652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6303c5cc-3340-45d2-a4ab-4500396246ef tempest-FloatingIPsAssociationTestJSON-1584539085 tempest-FloatingIPsAssociationTestJSON-1584539085-project-member] Lock "46435c68-f85a-4360-b2b5-6296afc33c3e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.807s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.287582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 870.287582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Fetch image to [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8/OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 870.287582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Downloading stream optimized image acef627a-6665-4737-891e-74c2350a5188 to [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8/OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8.vmdk on the data store datastore2 as vApp {{(pid=61964) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 870.287853] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Downloading image file data acef627a-6665-4737-891e-74c2350a5188 to the ESX as VM named 'OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8' {{(pid=61964) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 870.363858] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 870.363858] env[61964]: value = "resgroup-9" [ 870.363858] env[61964]: _type = "ResourcePool" [ 870.363858] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 870.364406] env[61964]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0c221b3e-b60a-45f3-b656-62498ec694ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.389265] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease: (returnval){ [ 870.389265] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d90df6-6d49-48d9-ef9f-3b8eadd0c28d" [ 870.389265] env[61964]: _type = "HttpNfcLease" [ 870.389265] env[61964]: } obtained for vApp import into resource pool (val){ [ 870.389265] env[61964]: value = "resgroup-9" [ 870.389265] env[61964]: _type = "ResourcePool" [ 870.389265] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 870.389564] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the lease: (returnval){ [ 870.389564] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d90df6-6d49-48d9-ef9f-3b8eadd0c28d" [ 870.389564] env[61964]: _type = "HttpNfcLease" [ 870.389564] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 870.397938] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.397938] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d90df6-6d49-48d9-ef9f-3b8eadd0c28d" [ 870.397938] env[61964]: _type = "HttpNfcLease" [ 870.397938] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 870.405890] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 870.405890] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.406664] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 870.406664] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.406664] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 870.406664] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 870.406823] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 870.406995] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 870.407194] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 870.407366] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 870.407549] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 870.417913] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7063334-ac1f-46aa-869e-9a5e061ef234 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.437477] env[61964]: DEBUG oslo_vmware.api [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041171, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14014} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.438822] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.439016] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.439246] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.441256] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 870.441256] env[61964]: value = "task-1041173" [ 870.441256] env[61964]: _type = "Task" [ 870.441256] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.452937] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041173, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.466149] env[61964]: INFO nova.scheduler.client.report [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted allocations for instance 63793ce6-d511-403e-8a4b-cad8c4157449 [ 870.540147] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71ab392-9818-403a-b00f-9167b26f3f0c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.549329] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa48e3e5-8a09-4aea-a64d-9c5c96191166 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.554367] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.583588] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c384bf-8ff0-4d12-bc45-a720c31e4d93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.592329] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9db593-cedb-4773-b5cc-6fe367977a87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.606763] env[61964]: DEBUG nova.compute.provider_tree [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.898699] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.898699] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d90df6-6d49-48d9-ef9f-3b8eadd0c28d" [ 870.898699] env[61964]: _type = "HttpNfcLease" [ 870.898699] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 870.899062] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 870.899062] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d90df6-6d49-48d9-ef9f-3b8eadd0c28d" [ 870.899062] env[61964]: _type = "HttpNfcLease" [ 870.899062] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 870.899855] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cdee2c-593d-46c4-b2b0-907db4d08c5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.908344] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 870.908566] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 870.975705] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.980381] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-55cabc86-a706-43c8-bff5-0d092a57710a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.982235] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041173, 'name': ReconfigVM_Task, 'duration_secs': 0.171601} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.983286] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 871.112981] env[61964]: DEBUG nova.scheduler.client.report [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.491101] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 871.491388] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.491566] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 871.491784] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.491998] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 871.492237] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 871.492488] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 871.492720] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 871.492936] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 871.493139] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 871.493901] env[61964]: DEBUG nova.virt.hardware [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 871.499449] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfiguring VM instance instance-0000002c to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 871.504061] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53c27545-9e94-4fe9-8731-8ce8c083e953 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.531008] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 871.531008] env[61964]: value = "task-1041174" [ 871.531008] env[61964]: _type = "Task" [ 871.531008] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.544972] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041174, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.623239] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.626070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.626392] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.626619] env[61964]: INFO nova.compute.manager [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Attaching volume 090cdd91-5f65-47f9-a5fb-3e8504507812 to /dev/sdb [ 871.628598] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.049s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.628888] env[61964]: DEBUG nova.objects.instance [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lazy-loading 'resources' on Instance uuid e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.647273] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 871.647574] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 871.649071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8134916-2df8-47c9-b2f3-f7a9037b7dc7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.657975] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 871.658164] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 871.658512] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4a09d758-bcd3-444c-a8ac-4339cf4e701e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.665563] env[61964]: INFO nova.scheduler.client.report [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Deleted allocations for instance 93720ce5-834a-4d8d-b230-13d38620b688 [ 871.671189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb829b86-4cdf-4510-9135-2abebc798ad3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.679555] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea2468f-41c9-404e-a78e-dbd677f717cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.700944] env[61964]: DEBUG nova.virt.block_device [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating existing volume attachment record: 73fc7a73-8071-4f7c-8eea-a4e9d9a8b991 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 871.881120] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529772fe-f946-ad62-6e4d-77f26c6322f0/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 871.881207] env[61964]: INFO nova.virt.vmwareapi.images [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Downloaded image file data acef627a-6665-4737-891e-74c2350a5188 [ 871.882082] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eaaafe-eaa6-4d4a-afc6-4e9e174679f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.898337] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95d804f4-585b-42fc-b61a-5a8c7a3c9504 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.924777] env[61964]: INFO nova.virt.vmwareapi.images [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] The imported VM was unregistered [ 871.927444] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 871.927686] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.928012] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e895bb2c-93bb-4c0f-830c-3f7d5fe50050 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.939146] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.939371] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8/OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8.vmdk to [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk. {{(pid=61964) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 871.939651] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a8337360-c6a3-4776-9c2b-71c193bc3f60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.948112] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 871.948112] env[61964]: value = "task-1041177" [ 871.948112] env[61964]: _type = "Task" [ 871.948112] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.960140] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.027967] env[61964]: DEBUG nova.compute.manager [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 872.027967] env[61964]: DEBUG nova.compute.manager [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing instance network info cache due to event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 872.028178] env[61964]: DEBUG oslo_concurrency.lockutils [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.028386] env[61964]: DEBUG oslo_concurrency.lockutils [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.028458] env[61964]: DEBUG nova.network.neutron [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.043544] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041174, 'name': ReconfigVM_Task, 'duration_secs': 0.191932} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.043840] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfigured VM instance instance-0000002c to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 872.044712] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e3aa41-3a97-4f60-98a1-8b22e8c91f3f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.075247] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.076410] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1278b64c-12bd-4992-aa7a-d0e4152b58e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.103383] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 872.103383] env[61964]: value = "task-1041179" [ 872.103383] env[61964]: _type = "Task" [ 872.103383] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.114715] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.181813] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b816354e-0c3f-4b3c-a883-14828b632054 tempest-ServerShowV257Test-786967685 tempest-ServerShowV257Test-786967685-project-member] Lock "93720ce5-834a-4d8d-b230-13d38620b688" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.490s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.448892] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912023fa-ed1f-48f5-9d79-c9f6fb9b1507 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.466286] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cd595c-1f7b-4f76-b4ca-8c408188c909 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.472741] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.504008] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9b5e74-190d-46e8-b9c4-c426e256bfdc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.513481] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97edf958-1105-4113-aba3-89f6e341f1f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.530084] env[61964]: DEBUG nova.compute.provider_tree [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.619536] env[61964]: DEBUG oslo_vmware.api [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041179, 'name': ReconfigVM_Task, 'duration_secs': 0.399821} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.619883] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Reconfigured VM instance instance-0000002c to attach disk [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4/df0a9a69-bd44-4da4-ba3a-9ba241c010a4.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.620200] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 872.801437] env[61964]: DEBUG nova.network.neutron [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updated VIF entry in instance network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.801841] env[61964]: DEBUG nova.network.neutron [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": null, "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapedda620f-12", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.963736] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.033594] env[61964]: DEBUG nova.scheduler.client.report [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.131711] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6464f2-93cd-4ad5-a516-3a29635c964b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.158880] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca97c8c-139f-483f-b112-d1a8e53fbb41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.178555] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 873.227312] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.304999] env[61964]: DEBUG oslo_concurrency.lockutils [req-7059564e-c71c-4cb3-9ed4-8091283d660d req-432d87ae-3db0-4698-80e3-3b2d92e20749 service nova] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.467365] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.541221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.912s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.544283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.562s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.544448] env[61964]: DEBUG nova.objects.instance [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lazy-loading 'resources' on Instance uuid eb97402e-e722-4cc3-a1d3-1b360ab1e1a7 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.573724] env[61964]: INFO nova.scheduler.client.report [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Deleted allocations for instance e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0 [ 873.725529] env[61964]: DEBUG nova.network.neutron [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Port 8b2c44a6-df10-43fa-9afa-85167edbb650 binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 873.965473] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.083689] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e0cddb94-de2c-4e58-b0bf-1127756e7939 tempest-ServerShowV254Test-440751991 tempest-ServerShowV254Test-440751991-project-member] Lock "e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.881s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.360159] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67af93bc-05f7-4384-896d-ba3cf57290d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.368563] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04f9fa9-ea71-41b2-bb3c-26a3c2d571e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.405024] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbb2eee-3a55-4742-bc39-12727c84cae6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.412645] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb2ce72-d9f4-40a5-87d2-4f3aef8ad6f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.428100] env[61964]: DEBUG nova.compute.provider_tree [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.465795] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041177, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.4319} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.465795] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8/OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8.vmdk to [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk. [ 874.465795] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Cleaning up location [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 874.466022] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_cc08b827-e6fd-4a9f-86ed-57814b4469a8 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.466259] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a9db31b-890e-42f0-b13a-96132d1233f5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.473702] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 874.473702] env[61964]: value = "task-1041182" [ 874.473702] env[61964]: _type = "Task" [ 874.473702] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.481760] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.760314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.760652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.760750] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.937762] env[61964]: DEBUG nova.scheduler.client.report [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 874.984578] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037957} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.984845] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.985035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.985306] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk to [datastore2] 50b04cbf-9064-4010-b350-638e8096bb96/50b04cbf-9064-4010-b350-638e8096bb96.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 874.985652] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f402b20-60ba-4f80-9b29-cf70d1ad34f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.993915] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 874.993915] env[61964]: value = "task-1041183" [ 874.993915] env[61964]: _type = "Task" [ 874.993915] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.003278] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.442303] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.444798] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.114s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.445017] env[61964]: DEBUG nova.objects.instance [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'resources' on Instance uuid e93f98b3-4763-4f02-abd3-c24a9ab8dc17 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.474698] env[61964]: INFO nova.scheduler.client.report [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Deleted allocations for instance eb97402e-e722-4cc3-a1d3-1b360ab1e1a7 [ 875.506505] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.829252] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.829252] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.829252] env[61964]: DEBUG nova.network.neutron [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.991909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-506dbfed-8266-452f-830d-bf9fc6cdbb6e tempest-ServerMetadataTestJSON-308897525 tempest-ServerMetadataTestJSON-308897525-project-member] Lock "eb97402e-e722-4cc3-a1d3-1b360ab1e1a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.074s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.014667] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.280776] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a0fe13-4f56-477a-85af-e2bcd890b56c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.290696] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd09f7c-feeb-4385-bae3-63af9d57aab6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.324189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855d34e9-a4a7-43cd-8112-658837448323 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.335514] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca86e4b-5b53-4a15-9551-e4163e255549 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.353335] env[61964]: DEBUG nova.compute.provider_tree [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.511694] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.766538] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 876.766538] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230511', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'name': 'volume-090cdd91-5f65-47f9-a5fb-3e8504507812', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '138f44e4-e12e-4f89-a9b2-8a512b53cdf5', 'attached_at': '', 'detached_at': '', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'serial': '090cdd91-5f65-47f9-a5fb-3e8504507812'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 876.770018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8892501-6c8b-4c85-af96-6d1107f15bf1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.792532] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be283b2a-69da-43b1-85d4-f28caa2504b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.820784] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] volume-090cdd91-5f65-47f9-a5fb-3e8504507812/volume-090cdd91-5f65-47f9-a5fb-3e8504507812.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.821638] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f517e2ac-d7e8-4315-a3ed-7fef0b044ddd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.843617] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 876.843617] env[61964]: value = "task-1041184" [ 876.843617] env[61964]: _type = "Task" [ 876.843617] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.855317] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041184, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.857097] env[61964]: DEBUG nova.scheduler.client.report [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.952672] env[61964]: DEBUG nova.network.neutron [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.012588] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.357507] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041184, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.362607] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.918s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.367023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.736s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.367023] env[61964]: DEBUG nova.objects.instance [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid fb96bf3a-ea2c-414d-bb29-eca2b07d51b6 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.394560] env[61964]: INFO nova.scheduler.client.report [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted allocations for instance e93f98b3-4763-4f02-abd3-c24a9ab8dc17 [ 877.455375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.511266] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041183, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.38107} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.511562] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/acef627a-6665-4737-891e-74c2350a5188/acef627a-6665-4737-891e-74c2350a5188.vmdk to [datastore2] 50b04cbf-9064-4010-b350-638e8096bb96/50b04cbf-9064-4010-b350-638e8096bb96.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 877.512421] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8d1e81-b69d-4e95-b19a-95ece04981f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.538927] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 50b04cbf-9064-4010-b350-638e8096bb96/50b04cbf-9064-4010-b350-638e8096bb96.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.539277] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a42e2ee2-c34c-41cf-9e59-cfe7a81b04d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.562827] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 877.562827] env[61964]: value = "task-1041185" [ 877.562827] env[61964]: _type = "Task" [ 877.562827] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.571924] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.855165] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041184, 'name': ReconfigVM_Task, 'duration_secs': 0.973338} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.855602] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfigured VM instance instance-00000044 to attach disk [datastore2] volume-090cdd91-5f65-47f9-a5fb-3e8504507812/volume-090cdd91-5f65-47f9-a5fb-3e8504507812.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.860941] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae4d0276-32f3-4ae9-80e9-887b1dfa98d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.880291] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 877.880291] env[61964]: value = "task-1041186" [ 877.880291] env[61964]: _type = "Task" [ 877.880291] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.898995] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041186, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.910091] env[61964]: DEBUG oslo_concurrency.lockutils [None req-711dd607-0f60-48b5-bbc2-3a953ca0e703 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "e93f98b3-4763-4f02-abd3-c24a9ab8dc17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.705s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.981719] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71748d0-9842-46c7-9e2b-0f2732b3c798 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.010036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4aa415-239e-4787-b17a-5570ab53033a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.019411] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 878.078288] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041185, 'name': ReconfigVM_Task, 'duration_secs': 0.301842} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.078288] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 50b04cbf-9064-4010-b350-638e8096bb96/50b04cbf-9064-4010-b350-638e8096bb96.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.078288] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3231b4f8-7095-45f5-b085-637d96c074f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.089455] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 878.089455] env[61964]: value = "task-1041187" [ 878.089455] env[61964]: _type = "Task" [ 878.089455] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.104757] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041187, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.228942] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edded938-a185-4477-8ba0-73fe62cf9931 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.238170] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a86d044-c2d4-4d14-a6b3-31f3def7876b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.279073] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7384a921-ad8d-47d9-b875-b015af9dd0fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.287642] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c57d5e4-2393-4c5b-af5a-9876b565afee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.303481] env[61964]: DEBUG nova.compute.provider_tree [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.397114] env[61964]: DEBUG oslo_vmware.api [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041186, 'name': ReconfigVM_Task, 'duration_secs': 0.208705} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.397399] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230511', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'name': 'volume-090cdd91-5f65-47f9-a5fb-3e8504507812', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '138f44e4-e12e-4f89-a9b2-8a512b53cdf5', 'attached_at': '', 'detached_at': '', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'serial': '090cdd91-5f65-47f9-a5fb-3e8504507812'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 878.527023] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b2be60cd-f8d8-492c-afda-1443e5583cfd tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance 'df0a9a69-bd44-4da4-ba3a-9ba241c010a4' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 878.603950] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041187, 'name': Rename_Task, 'duration_secs': 0.162567} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.604363] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 878.604629] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de50d7fc-b727-4f24-9658-f4e265c4c7bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.613637] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 878.613637] env[61964]: value = "task-1041188" [ 878.613637] env[61964]: _type = "Task" [ 878.613637] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.626069] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.806875] env[61964]: DEBUG nova.scheduler.client.report [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.125931] env[61964]: DEBUG oslo_vmware.api [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041188, 'name': PowerOnVM_Task, 'duration_secs': 0.485012} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.126228] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 879.126425] env[61964]: INFO nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Took 13.70 seconds to spawn the instance on the hypervisor. [ 879.126650] env[61964]: DEBUG nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.127466] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d56b85f-83f4-4ff7-85cd-ac26dff829c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.313415] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.948s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.317218] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.671s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.317218] env[61964]: DEBUG nova.objects.instance [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lazy-loading 'resources' on Instance uuid 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.353337] env[61964]: INFO nova.scheduler.client.report [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance fb96bf3a-ea2c-414d-bb29-eca2b07d51b6 [ 879.468815] env[61964]: DEBUG nova.objects.instance [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.649358] env[61964]: INFO nova.compute.manager [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Took 33.82 seconds to build instance. [ 879.698957] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.699234] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.870030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-21e28298-9b87-49bd-b18c-4c4807a3239a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "fb96bf3a-ea2c-414d-bb29-eca2b07d51b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.729s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.972869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d159f9f7-7789-4233-a067-775c12de51c0 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.346s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.083016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c859400e-1cbc-4697-8c04-26a22fa65677 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.092143] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fc294e-1ed3-48f3-ae01-ef13ec3438ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.136260] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1acf9e3-508e-4b9c-8e80-52f5ff7b4b1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.149739] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac5226f-ab45-470e-8d19-2ee5375540f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.153387] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0666df0-ec1d-429b-8b45-8e81c59f1477 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.338s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.164196] env[61964]: DEBUG nova.compute.provider_tree [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.202301] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 880.225458] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.225458] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.225458] env[61964]: DEBUG nova.compute.manager [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 880.226064] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687ee0ff-d110-4ad8-8bbb-1ea75162ce08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.235842] env[61964]: DEBUG nova.compute.manager [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 880.236436] env[61964]: DEBUG nova.objects.instance [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.671024] env[61964]: DEBUG nova.scheduler.client.report [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.729174] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.867513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "4d9c24cf-e11b-4d95-ad41-538d91279689" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.867513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.174139] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.857s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.177484] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.446s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.179551] env[61964]: INFO nova.compute.claims [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.193770] env[61964]: INFO nova.scheduler.client.report [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Deleted allocations for instance 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4 [ 881.234947] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "50b04cbf-9064-4010-b350-638e8096bb96" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.235219] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.235425] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "50b04cbf-9064-4010-b350-638e8096bb96-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.236627] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.236627] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.243404] env[61964]: INFO nova.compute.manager [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Terminating instance [ 881.249384] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 881.250032] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ea7b7c0-8fe5-46a9-994c-431c643e30f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.259094] env[61964]: DEBUG oslo_vmware.api [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 881.259094] env[61964]: value = "task-1041189" [ 881.259094] env[61964]: _type = "Task" [ 881.259094] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.271102] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.271360] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.272815] env[61964]: DEBUG oslo_vmware.api [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.369121] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 881.648384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.648658] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.702352] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9bd15577-00a7-4831-83b4-9d66591bea60 tempest-SecurityGroupsTestJSON-1215864552 tempest-SecurityGroupsTestJSON-1215864552-project-member] Lock "6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.780s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.734277] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.734523] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.734725] env[61964]: DEBUG nova.compute.manager [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Going to confirm migration 2 {{(pid=61964) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5218}} [ 881.747687] env[61964]: DEBUG nova.compute.manager [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 881.748179] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 881.749165] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f064983e-9c73-4a8c-a53a-a2fcf609932b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.759511] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 881.759848] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f48bcde4-616b-4f14-9349-8c551c375ede {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.771800] env[61964]: DEBUG oslo_vmware.api [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041189, 'name': PowerOffVM_Task, 'duration_secs': 0.268831} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.772992] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 881.773234] env[61964]: DEBUG nova.compute.manager [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 881.773924] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 881.773924] env[61964]: value = "task-1041190" [ 881.773924] env[61964]: _type = "Task" [ 881.773924] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.774641] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 881.777778] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f05a4e-a95a-4c0b-b0dc-265e607d591a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.793351] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.893819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.151145] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 882.299463] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041190, 'name': PowerOffVM_Task, 'duration_secs': 0.230072} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.299800] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 882.299980] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 882.302074] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df89e547-8632-4829-b074-e8547f07fb9e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.304729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b06ade99-1019-455b-a157-d4f93ecfd678 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.079s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.306035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.331493] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.331493] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.331718] env[61964]: DEBUG nova.network.neutron [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.332043] env[61964]: DEBUG nova.objects.instance [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'info_cache' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.400035] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 882.400035] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 882.400219] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore2] 50b04cbf-9064-4010-b350-638e8096bb96 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.400491] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a4aa2fe-b802-42f3-9776-59f49fdc5c2a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.408719] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 882.408719] env[61964]: value = "task-1041192" [ 882.408719] env[61964]: _type = "Task" [ 882.408719] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.420164] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.476505] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c904eae-5541-4fdf-982c-8837ff5e7e27 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.484297] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e703475d-91bd-4b5e-bb23-17caf5f4e0c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.516853] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b00331f-a014-41b4-ab7b-2ebe797499ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.526144] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f17b058-d1b1-4b5e-87ce-613e038bdcfb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.544017] env[61964]: DEBUG nova.compute.provider_tree [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.674389] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.919809] env[61964]: DEBUG oslo_vmware.api [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389594} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.920104] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.920555] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 882.920555] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 882.920671] env[61964]: INFO nova.compute.manager [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Took 1.17 seconds to destroy the instance on the hypervisor. [ 882.920882] env[61964]: DEBUG oslo.service.loopingcall [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.921091] env[61964]: DEBUG nova.compute.manager [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 882.922225] env[61964]: DEBUG nova.network.neutron [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 882.944143] env[61964]: DEBUG nova.objects.instance [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.047245] env[61964]: DEBUG nova.scheduler.client.report [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.334166] env[61964]: DEBUG nova.compute.manager [req-5adbb1a6-167a-4c35-8af9-01c70bd5dbbd req-d456de8d-a7a6-48ac-a9ed-5eb7faac49c2 service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Received event network-vif-deleted-31b39082-5123-4292-bb5a-18184dc24861 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 883.334166] env[61964]: INFO nova.compute.manager [req-5adbb1a6-167a-4c35-8af9-01c70bd5dbbd req-d456de8d-a7a6-48ac-a9ed-5eb7faac49c2 service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Neutron deleted interface 31b39082-5123-4292-bb5a-18184dc24861; detaching it from the instance and deleting it from the info cache [ 883.334166] env[61964]: DEBUG nova.network.neutron [req-5adbb1a6-167a-4c35-8af9-01c70bd5dbbd req-d456de8d-a7a6-48ac-a9ed-5eb7faac49c2 service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.454354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.454553] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.454794] env[61964]: DEBUG nova.network.neutron [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.455060] env[61964]: DEBUG nova.objects.instance [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'info_cache' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.552044] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.554236] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 883.564050] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.106s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.611264] env[61964]: DEBUG nova.network.neutron [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.816963] env[61964]: DEBUG nova.network.neutron [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.837291] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd9759fb-16f2-4580-9d93-2e3748a734bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.849255] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4042e6b-2940-42dd-b86f-cfb2adab0601 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.889916] env[61964]: DEBUG nova.compute.manager [req-5adbb1a6-167a-4c35-8af9-01c70bd5dbbd req-d456de8d-a7a6-48ac-a9ed-5eb7faac49c2 service nova] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Detach interface failed, port_id=31b39082-5123-4292-bb5a-18184dc24861, reason: Instance 50b04cbf-9064-4010-b350-638e8096bb96 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 883.959014] env[61964]: DEBUG nova.objects.base [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Object Instance<138f44e4-e12e-4f89-a9b2-8a512b53cdf5> lazy-loaded attributes: flavor,info_cache {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 884.064689] env[61964]: DEBUG nova.compute.utils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.066151] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 884.066282] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.070735] env[61964]: INFO nova.compute.claims [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.115058] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.115332] env[61964]: DEBUG nova.objects.instance [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'migration_context' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.118380] env[61964]: DEBUG nova.policy [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.320885] env[61964]: INFO nova.compute.manager [-] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Took 1.40 seconds to deallocate network for instance. [ 884.403224] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Successfully created port: 9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.575524] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 884.579130] env[61964]: INFO nova.compute.resource_tracker [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating resource usage from migration 8b2cfb5a-d6e6-49e9-8d5f-74cdad8f363a [ 884.618089] env[61964]: DEBUG nova.objects.base [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 884.618878] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945f0eb3-e4a7-4ecd-9e13-c4fff3badc0b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.646422] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-939adedf-c2f6-4fbb-b10e-7c654597cbbd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.653077] env[61964]: DEBUG oslo_vmware.api [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 884.653077] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c7555f-2351-0d85-896d-662249abb9f7" [ 884.653077] env[61964]: _type = "Task" [ 884.653077] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.660604] env[61964]: DEBUG oslo_vmware.api [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c7555f-2351-0d85-896d-662249abb9f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.827029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.868749] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d15b000-a4a2-4a39-968e-e18ccf9881c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.876986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed01975-2298-4903-bcdb-d2c1237abecd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.885228] env[61964]: DEBUG nova.network.neutron [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.924154] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24abd4e-608f-4a20-8d97-5da02a2174ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.932099] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493a8178-b5ca-4508-8f20-7085981ef600 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.946781] env[61964]: DEBUG nova.compute.provider_tree [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.163956] env[61964]: DEBUG oslo_vmware.api [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c7555f-2351-0d85-896d-662249abb9f7, 'name': SearchDatastore_Task, 'duration_secs': 0.007385} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.164126] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.389217] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.451263] env[61964]: DEBUG nova.scheduler.client.report [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 885.591801] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 885.613133] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 885.613396] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.613555] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.613740] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.613933] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.614056] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 885.614270] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 885.614435] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 885.614624] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 885.614795] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 885.614972] env[61964]: DEBUG nova.virt.hardware [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 885.615840] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f642df6-ea07-4322-8db4-cc4df7a54cfe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.623726] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e28b4d-1b29-48c6-86c6-ddf8339e150d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.960298] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.399s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.960526] env[61964]: INFO nova.compute.manager [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Migrating [ 885.967053] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.518s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.967267] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.967403] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 885.967704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.974s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.968303] env[61964]: DEBUG nova.objects.instance [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lazy-loading 'resources' on Instance uuid a1e5c369-4233-4794-adda-1ba1e89e8154 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.969939] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be6e2f2-573a-411c-bd4a-5cc9b487308f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.987016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23a4cdf-72ad-432e-9150-715386609ba5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.004307] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b571d083-6daf-4e79-b7fa-dbbdcc5aece4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.014269] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313073d9-3b24-464f-8c68-fab4759c3f6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.044468] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179288MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 886.044820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.265614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "f5acabe4-674b-4c4c-85b4-b9270926017b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.265932] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.399222] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 886.399654] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1a240cd-a803-4e08-b5c0-15ae388afc7f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.407671] env[61964]: DEBUG oslo_vmware.api [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 886.407671] env[61964]: value = "task-1041193" [ 886.407671] env[61964]: _type = "Task" [ 886.407671] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.417459] env[61964]: DEBUG oslo_vmware.api [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.430332] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Successfully updated port: 9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.482609] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.482873] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.483055] env[61964]: DEBUG nova.network.neutron [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.503031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.503031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.563310] env[61964]: DEBUG nova.compute.manager [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received event network-vif-plugged-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 886.563662] env[61964]: DEBUG oslo_concurrency.lockutils [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.563923] env[61964]: DEBUG oslo_concurrency.lockutils [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.564135] env[61964]: DEBUG oslo_concurrency.lockutils [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.564769] env[61964]: DEBUG nova.compute.manager [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] No waiting events found dispatching network-vif-plugged-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 886.564769] env[61964]: WARNING nova.compute.manager [req-74c1e392-6926-4896-b54e-e06f38968ca2 req-50342551-62b2-4492-9017-ae3c33b43a2f service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received unexpected event network-vif-plugged-9e14aef8-cd53-45ad-9b3c-152397bc01da for instance with vm_state building and task_state spawning. [ 886.740288] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823ee095-744b-4dff-9273-18072c3d30bf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.747926] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf66d496-a003-4692-9316-14012e5fa0e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.783705] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 886.787496] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c3d4dd-3134-4273-bb5d-f68996714659 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.795793] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cd91db-b9e2-4b3f-9ad1-091f3ab13fbb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.809145] env[61964]: DEBUG nova.compute.provider_tree [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.896446] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "d8f24177-b695-4d40-958d-9272b4f683cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.896686] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.924205] env[61964]: DEBUG oslo_vmware.api [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041193, 'name': PowerOnVM_Task, 'duration_secs': 0.391118} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.924205] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.924205] env[61964]: DEBUG nova.compute.manager [None req-8881182e-da1e-4752-bb08-e0b5b93e5981 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.925537] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a78ddf0-101f-47bc-82ab-c5e73ee9eb5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.934485] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.934623] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.935326] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.009640] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 887.236576] env[61964]: DEBUG nova.network.neutron [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [{"id": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "address": "fa:16:3e:30:78:4d", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a884ac5-73", "ovs_interfaceid": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.308248] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.312530] env[61964]: DEBUG nova.scheduler.client.report [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.399219] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 887.470087] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.524835] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.661853] env[61964]: DEBUG nova.network.neutron [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updating instance_info_cache with network_info: [{"id": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "address": "fa:16:3e:b8:3b:74", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e14aef8-cd", "ovs_interfaceid": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.739438] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.817241] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.819563] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.709s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.821386] env[61964]: INFO nova.compute.claims [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.833978] env[61964]: INFO nova.scheduler.client.report [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Deleted allocations for instance a1e5c369-4233-4794-adda-1ba1e89e8154 [ 887.917844] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.165429] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.165783] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance network_info: |[{"id": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "address": "fa:16:3e:b8:3b:74", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e14aef8-cd", "ovs_interfaceid": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 888.166269] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:3b:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e14aef8-cd53-45ad-9b3c-152397bc01da', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.173814] env[61964]: DEBUG oslo.service.loopingcall [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.174047] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.174284] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8159309a-ce90-4e49-b141-9b8649cdfc42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.193818] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.193818] env[61964]: value = "task-1041194" [ 888.193818] env[61964]: _type = "Task" [ 888.193818] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.201372] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041194, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.340925] env[61964]: DEBUG oslo_concurrency.lockutils [None req-de5003f9-f940-43f6-82da-9fefc5703eed tempest-ServerAddressesNegativeTestJSON-1170520902 tempest-ServerAddressesNegativeTestJSON-1170520902-project-member] Lock "a1e5c369-4233-4794-adda-1ba1e89e8154" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.910s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.588131] env[61964]: DEBUG nova.compute.manager [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received event network-changed-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 888.588131] env[61964]: DEBUG nova.compute.manager [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Refreshing instance network info cache due to event network-changed-9e14aef8-cd53-45ad-9b3c-152397bc01da. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 888.588131] env[61964]: DEBUG oslo_concurrency.lockutils [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] Acquiring lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.588131] env[61964]: DEBUG oslo_concurrency.lockutils [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] Acquired lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.588131] env[61964]: DEBUG nova.network.neutron [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Refreshing network info cache for port 9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.707125] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041194, 'name': CreateVM_Task, 'duration_secs': 0.309291} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.707392] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 888.707983] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.708197] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.708526] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.708783] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0278977-16d2-4069-bb14-b9717e5e46aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.713404] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 888.713404] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226be8b-44cb-5525-086b-4211d596c118" [ 888.713404] env[61964]: _type = "Task" [ 888.713404] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.724227] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226be8b-44cb-5525-086b-4211d596c118, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.113077] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2e4b38-7d7a-405e-8af5-ac91e6696463 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.119170] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a2350c-34a0-4144-9394-ed9c62fbff69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.152800] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc608016-64b4-4a85-b47f-aa22cc04baf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.165222] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fff310-0c98-4115-963c-05eb9953fe97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.175726] env[61964]: DEBUG nova.compute.provider_tree [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.225727] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5226be8b-44cb-5525-086b-4211d596c118, 'name': SearchDatastore_Task, 'duration_secs': 0.008799} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.226311] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.226808] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.227225] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.227513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.227827] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.228260] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b182c4d-c9aa-4e17-83cb-c0fef0747a63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.239018] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.239018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 889.239018] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-231a1aa7-eb48-48b1-9431-0d7a4f3b0738 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.246017] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 889.246017] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524e8dcf-d56d-1417-5f37-943de4511d4d" [ 889.246017] env[61964]: _type = "Task" [ 889.246017] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.254754] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524e8dcf-d56d-1417-5f37-943de4511d4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.258331] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a214a394-a4ec-45e7-8699-372ab84bd315 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.277157] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 889.509475] env[61964]: DEBUG nova.network.neutron [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updated VIF entry in instance network info cache for port 9e14aef8-cd53-45ad-9b3c-152397bc01da. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.509921] env[61964]: DEBUG nova.network.neutron [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updating instance_info_cache with network_info: [{"id": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "address": "fa:16:3e:b8:3b:74", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e14aef8-cd", "ovs_interfaceid": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.679579] env[61964]: DEBUG nova.scheduler.client.report [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.754431] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524e8dcf-d56d-1417-5f37-943de4511d4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008261} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.755226] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca931c1d-a931-4d5e-bd71-b4eecb78ca79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.760545] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 889.760545] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5291c1db-e4e2-96b6-176c-729983f9a791" [ 889.760545] env[61964]: _type = "Task" [ 889.760545] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.768048] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5291c1db-e4e2-96b6-176c-729983f9a791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.782998] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.783218] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e43f3520-9ba2-4668-9976-5d6af7ab8ccd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.790608] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 889.790608] env[61964]: value = "task-1041195" [ 889.790608] env[61964]: _type = "Task" [ 889.790608] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.798232] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.012367] env[61964]: DEBUG oslo_concurrency.lockutils [req-f8f9369f-e0d0-4c89-bbe5-6a92cc6c57ac req-12feaf73-daa9-41b5-a62e-158ee8323abd service nova] Releasing lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.186982] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.187474] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 890.190407] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.835s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.192024] env[61964]: INFO nova.compute.claims [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.270977] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5291c1db-e4e2-96b6-176c-729983f9a791, 'name': SearchDatastore_Task, 'duration_secs': 0.008898} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.271533] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.271682] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f73b99bf-6a2b-4f21-b855-74965c95ed76/f73b99bf-6a2b-4f21-b855-74965c95ed76.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 890.271844] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1f025a2-1f04-44ab-ab63-59caeec1065c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.278133] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 890.278133] env[61964]: value = "task-1041196" [ 890.278133] env[61964]: _type = "Task" [ 890.278133] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.285498] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.298230] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041195, 'name': PowerOffVM_Task, 'duration_secs': 0.451392} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.298489] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.298674] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 890.697106] env[61964]: DEBUG nova.compute.utils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.700959] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 890.787881] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450164} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.789223] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f73b99bf-6a2b-4f21-b855-74965c95ed76/f73b99bf-6a2b-4f21-b855-74965c95ed76.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 890.789223] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.789223] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dd18199-6ca0-4250-aa95-b864c62cac09 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.795265] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 890.795265] env[61964]: value = "task-1041197" [ 890.795265] env[61964]: _type = "Task" [ 890.795265] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.802351] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041197, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.804382] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 890.804645] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.804842] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 890.805080] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.805251] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 890.805402] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 890.805624] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 890.805833] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 890.806036] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 890.806216] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 890.806395] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 890.811334] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-175c7b3b-4998-4c28-a396-3b9e3e564b05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.826032] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 890.826032] env[61964]: value = "task-1041198" [ 890.826032] env[61964]: _type = "Task" [ 890.826032] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.834175] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.201589] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 891.307287] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041197, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091476} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.307582] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.308363] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2987f17b-6e67-45d2-9add-4f6a9e0981c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.332530] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] f73b99bf-6a2b-4f21-b855-74965c95ed76/f73b99bf-6a2b-4f21-b855-74965c95ed76.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.335009] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9322991-5341-426b-ac4f-5de7d9ab84c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.356571] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041198, 'name': ReconfigVM_Task, 'duration_secs': 0.387658} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.357905] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 891.361170] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 891.361170] env[61964]: value = "task-1041199" [ 891.361170] env[61964]: _type = "Task" [ 891.361170] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.372123] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041199, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.478602] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac6dc61-1f24-4248-be75-8213bf9073fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.486618] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de239ca-3686-47f1-9838-67639eb93bd5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.518321] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85891051-9c46-4fc7-8f53-db00baa38800 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.525498] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19df1d08-d48c-4e20-a4a6-647d0cd4afbd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.539403] env[61964]: DEBUG nova.compute.provider_tree [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.866009] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 891.866325] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.866370] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 891.866560] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.866715] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 891.866866] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 891.867084] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 891.867252] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 891.867423] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 891.867590] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 891.867767] env[61964]: DEBUG nova.virt.hardware [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 891.873109] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfiguring VM instance instance-0000004a to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.873765] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd19e6ed-8c6c-4be5-b74c-24b78b757fa8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.896008] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041199, 'name': ReconfigVM_Task, 'duration_secs': 0.291091} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.897198] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Reconfigured VM instance instance-0000004d to attach disk [datastore2] f73b99bf-6a2b-4f21-b855-74965c95ed76/f73b99bf-6a2b-4f21-b855-74965c95ed76.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.897881] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 891.897881] env[61964]: value = "task-1041200" [ 891.897881] env[61964]: _type = "Task" [ 891.897881] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.898091] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e19b0909-ff1a-4540-a79a-44eebe285518 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.906928] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041200, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.908042] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 891.908042] env[61964]: value = "task-1041201" [ 891.908042] env[61964]: _type = "Task" [ 891.908042] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.916007] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041201, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.046301] env[61964]: DEBUG nova.scheduler.client.report [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 892.214600] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 892.243699] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 892.243947] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.244126] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 892.244317] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.244469] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 892.244620] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 892.244831] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 892.244996] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 892.245183] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 892.245351] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 892.245526] env[61964]: DEBUG nova.virt.hardware [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 892.246409] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d736959c-d795-4177-bb89-0bce4eb7f853 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.254322] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdceb0e1-3f7b-4327-80aa-9a06a446816e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.267215] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.272631] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Creating folder: Project (2a938ac2e12d4887a5a5c949c6272ba4). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 892.272878] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b56e990-3c50-4bc0-a1f3-b1ee8b2f80d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.281494] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Created folder: Project (2a938ac2e12d4887a5a5c949c6272ba4) in parent group-v230360. [ 892.281678] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Creating folder: Instances. Parent ref: group-v230513. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 892.281900] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a81c708c-be98-420d-a4bf-3b5b9e3eeb12 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.288687] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Created folder: Instances in parent group-v230513. [ 892.288887] env[61964]: DEBUG oslo.service.loopingcall [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.289114] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 892.289302] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a116762c-27d8-4e4a-ab53-9d9c58de7f0b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.305342] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.305342] env[61964]: value = "task-1041204" [ 892.305342] env[61964]: _type = "Task" [ 892.305342] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.311728] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041204, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.409061] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041200, 'name': ReconfigVM_Task, 'duration_secs': 0.198282} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.411884] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfigured VM instance instance-0000004a to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 892.412596] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae44359-10e9-49b9-8dcd-e0a13f48edf0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.419400] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041201, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.436500] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.436751] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e732b97-2070-46d4-99c6-bc0641904b4f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.453859] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 892.453859] env[61964]: value = "task-1041205" [ 892.453859] env[61964]: _type = "Task" [ 892.453859] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.460963] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041205, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.560368] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.560945] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 892.563909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.710s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.564147] env[61964]: DEBUG nova.objects.instance [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lazy-loading 'resources' on Instance uuid a995ff8c-75b9-4089-ad4c-9e6baa33fe8c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.816065] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041204, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.918721] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041201, 'name': Rename_Task, 'duration_secs': 0.833219} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.919009] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 892.919270] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95e5a84b-c9b0-41e0-8c01-b225d8b881ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.926095] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 892.926095] env[61964]: value = "task-1041206" [ 892.926095] env[61964]: _type = "Task" [ 892.926095] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.933828] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.966580] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041205, 'name': ReconfigVM_Task, 'duration_secs': 0.313277} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.966879] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f/84821045-27b2-4a99-87f9-988b8615b83f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.967190] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 893.068266] env[61964]: DEBUG nova.compute.utils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.075226] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Not allocating networking since 'none' was specified. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 893.318870] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041204, 'name': CreateVM_Task, 'duration_secs': 0.620226} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.318870] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 893.321638] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.321814] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.322236] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.323688] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c90f73-021f-4a5f-9c83-93ec5e40caad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.327928] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 893.327928] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c69fae-da4d-7bf9-d694-26f8f515d5f5" [ 893.327928] env[61964]: _type = "Task" [ 893.327928] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.340152] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c69fae-da4d-7bf9-d694-26f8f515d5f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.350944] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2bee0f1-9d72-49d0-817f-3908c27f4807 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.358408] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853ff1f2-2c96-4660-8f5c-bd51c26e5d8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.389722] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd902ae-157f-4a79-9cd3-e41adffd96ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.398904] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def46773-22d2-47fa-a58a-b22377fea1dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.419404] env[61964]: DEBUG nova.compute.provider_tree [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.443675] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041206, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.473997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d66fa7a-b64d-4790-bfca-d23042b5d3ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.494576] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f350b758-0f97-4ee4-b8aa-0588a5643684 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.515444] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 893.576025] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 893.842599] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c69fae-da4d-7bf9-d694-26f8f515d5f5, 'name': SearchDatastore_Task, 'duration_secs': 0.018541} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.843122] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.843558] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.843975] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.844282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.844626] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.845060] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c894554-c59e-40b1-8afb-2bb2d616c749 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.857943] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.858256] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 893.859378] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e97e1474-f86b-46f3-ae0f-7fd628f47d92 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.865944] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 893.865944] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52127a00-ee87-0c32-4d81-2c48bc5dd4fe" [ 893.865944] env[61964]: _type = "Task" [ 893.865944] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.874713] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52127a00-ee87-0c32-4d81-2c48bc5dd4fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.923047] env[61964]: DEBUG nova.scheduler.client.report [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.938333] env[61964]: DEBUG oslo_vmware.api [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041206, 'name': PowerOnVM_Task, 'duration_secs': 0.812653} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.939256] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 893.939467] env[61964]: INFO nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Took 8.35 seconds to spawn the instance on the hypervisor. [ 893.939650] env[61964]: DEBUG nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 893.940454] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab77946-8c4a-445a-bfdc-a912e55c09f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.057362] env[61964]: DEBUG nova.network.neutron [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Port 2a884ac5-7364-4168-bf0a-d947f84b8b92 binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 894.378287] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52127a00-ee87-0c32-4d81-2c48bc5dd4fe, 'name': SearchDatastore_Task, 'duration_secs': 0.015073} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.379128] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a3cd993-b279-4128-827b-a89c366ee43f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.384927] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 894.384927] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abf1a-fc56-e024-f4b1-a66c44a13d57" [ 894.384927] env[61964]: _type = "Task" [ 894.384927] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.392624] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abf1a-fc56-e024-f4b1-a66c44a13d57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.428902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.431652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.456s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.431922] env[61964]: DEBUG nova.objects.instance [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'resources' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.457988] env[61964]: INFO nova.scheduler.client.report [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Deleted allocations for instance a995ff8c-75b9-4089-ad4c-9e6baa33fe8c [ 894.462693] env[61964]: INFO nova.compute.manager [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Took 33.75 seconds to build instance. [ 894.584927] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 894.610139] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 894.612151] env[61964]: DEBUG nova.virt.hardware [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 894.612897] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b092b8d-28be-462e-bd9e-04a89d2a4870 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.621714] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6792563-48af-405e-ab25-6511e0e9abca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.637529] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.642909] env[61964]: DEBUG oslo.service.loopingcall [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.643414] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 894.643632] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d917920c-28b6-46ac-84e2-84f985ba1b9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.659703] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.659703] env[61964]: value = "task-1041207" [ 894.659703] env[61964]: _type = "Task" [ 894.659703] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.667241] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041207, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.895881] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523abf1a-fc56-e024-f4b1-a66c44a13d57, 'name': SearchDatastore_Task, 'duration_secs': 0.01001} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.896216] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.896494] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 52f358e4-beb2-4b96-8f96-5cd4da103eda/52f358e4-beb2-4b96-8f96-5cd4da103eda.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 894.896768] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2305a0f9-6f40-4d05-93c7-b487f789c7de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.904268] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 894.904268] env[61964]: value = "task-1041208" [ 894.904268] env[61964]: _type = "Task" [ 894.904268] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.913652] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041208, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.934720] env[61964]: DEBUG nova.objects.instance [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'numa_topology' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.967074] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2fdb60e2-b569-43b0-b43d-bbebf7e7c948 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.266s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.967773] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b259220-9037-48f4-a40c-743ff774c656 tempest-VolumesAdminNegativeTest-1258987589 tempest-VolumesAdminNegativeTest-1258987589-project-member] Lock "a995ff8c-75b9-4089-ad4c-9e6baa33fe8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.925s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.082440] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.082721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.082973] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.170748] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041207, 'name': CreateVM_Task, 'duration_secs': 0.258632} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.170925] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 895.171395] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.171565] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.171909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.172228] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1628bc88-2c89-45a4-90e8-29848a44fcc8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.177600] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 895.177600] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46ec4-e166-2e59-dbe5-b7cd66bf6b5b" [ 895.177600] env[61964]: _type = "Task" [ 895.177600] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.185244] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46ec4-e166-2e59-dbe5-b7cd66bf6b5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.275386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.275676] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.275869] env[61964]: INFO nova.compute.manager [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Shelving [ 895.414585] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041208, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.429012} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.414918] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 52f358e4-beb2-4b96-8f96-5cd4da103eda/52f358e4-beb2-4b96-8f96-5cd4da103eda.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 895.415162] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.415425] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22ad878a-2a12-4c36-b9d8-3a2b2e6341fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.422097] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 895.422097] env[61964]: value = "task-1041209" [ 895.422097] env[61964]: _type = "Task" [ 895.422097] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.429765] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041209, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.437369] env[61964]: DEBUG nova.objects.base [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Object Instance<63793ce6-d511-403e-8a4b-cad8c4157449> lazy-loaded attributes: resources,numa_topology {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 895.666808] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d040d2d2-6405-4804-9b5c-930bfb9ba8b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.674386] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9af7769-31ed-4412-8d00-45c626e61a31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.686291] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c46ec4-e166-2e59-dbe5-b7cd66bf6b5b, 'name': SearchDatastore_Task, 'duration_secs': 0.047882} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.709659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.709930] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.710194] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.710344] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.710524] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.710984] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1807213-0a00-40e8-ba16-57d48b656d5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.713366] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36523231-6731-4e90-b49b-1e1d90e5c553 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.720923] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36a09ed-7c7a-4089-a3ce-45898d308165 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.726097] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.726281] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.734047] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa99e23-46cd-4042-9028-80e44e67214e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.736549] env[61964]: DEBUG nova.compute.provider_tree [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.740186] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 895.740186] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524b447f-7945-a5ed-beba-137ab9d9db50" [ 895.740186] env[61964]: _type = "Task" [ 895.740186] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.748560] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524b447f-7945-a5ed-beba-137ab9d9db50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.932439] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041209, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065934} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.932705] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.933517] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac23bb92-1909-460a-b5a6-c4b334d33067 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.952369] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 52f358e4-beb2-4b96-8f96-5cd4da103eda/52f358e4-beb2-4b96-8f96-5cd4da103eda.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.952659] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58201109-a19e-4a19-bab9-0265f073f4e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.971823] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 895.971823] env[61964]: value = "task-1041210" [ 895.971823] env[61964]: _type = "Task" [ 895.971823] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.979303] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.134642] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.134845] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.135042] env[61964]: DEBUG nova.network.neutron [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.240134] env[61964]: DEBUG nova.scheduler.client.report [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.253995] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524b447f-7945-a5ed-beba-137ab9d9db50, 'name': SearchDatastore_Task, 'duration_secs': 0.04219} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.255376] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef6b1213-f1a2-4c55-ac45-8f5c72d4cb07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.261197] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 896.261197] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5237e184-d8df-431c-678b-c0f49978825b" [ 896.261197] env[61964]: _type = "Task" [ 896.261197] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.269344] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5237e184-d8df-431c-678b-c0f49978825b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.288511] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 896.288826] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe426683-4963-4426-b542-7314731dd225 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.295831] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 896.295831] env[61964]: value = "task-1041211" [ 896.295831] env[61964]: _type = "Task" [ 896.295831] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.303734] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041211, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.481945] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041210, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.748986] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.317s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.752370] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.023s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.753456] env[61964]: INFO nova.compute.claims [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.774621] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5237e184-d8df-431c-678b-c0f49978825b, 'name': SearchDatastore_Task, 'duration_secs': 0.010679} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.777558] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.777856] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 896.778486] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a03bd117-2450-4699-b9e9-e0643923c662 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.786501] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 896.786501] env[61964]: value = "task-1041212" [ 896.786501] env[61964]: _type = "Task" [ 896.786501] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.798241] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.809762] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041211, 'name': PowerOffVM_Task, 'duration_secs': 0.199764} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.810088] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 896.810911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d441670-55f8-423f-b794-98df3aa5867c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.830632] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09d9ad2-356d-4fd0-8ca5-79cee2e06247 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.984447] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041210, 'name': ReconfigVM_Task, 'duration_secs': 0.547963} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.984447] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 52f358e4-beb2-4b96-8f96-5cd4da103eda/52f358e4-beb2-4b96-8f96-5cd4da103eda.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.984447] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-743b18c5-34a9-4bf3-801c-8f110fe561b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.990498] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 896.990498] env[61964]: value = "task-1041213" [ 896.990498] env[61964]: _type = "Task" [ 896.990498] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.999926] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041213, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.001404] env[61964]: DEBUG nova.network.neutron [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [{"id": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "address": "fa:16:3e:30:78:4d", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a884ac5-73", "ovs_interfaceid": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.264681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-baf7080e-69ba-48ef-bccb-da174963999d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 46.955s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.267813] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 24.039s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.267813] env[61964]: INFO nova.compute.manager [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Unshelving [ 897.302000] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041212, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.343444] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.344186] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f806c986-58f4-47c6-979b-500a65030af5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.351112] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 897.351112] env[61964]: value = "task-1041214" [ 897.351112] env[61964]: _type = "Task" [ 897.351112] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.359171] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041214, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.501669] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041213, 'name': Rename_Task, 'duration_secs': 0.408242} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.501967] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 897.502484] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e5a25ba-8455-4cbe-8b6f-ec8c43b89fc8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.504438] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.516020] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 897.516020] env[61964]: value = "task-1041215" [ 897.516020] env[61964]: _type = "Task" [ 897.516020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.522057] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.802741] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55276} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.806149] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 897.806465] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.806963] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6a1710b-4e27-409c-9c1d-fa4b96945e4e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.818097] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 897.818097] env[61964]: value = "task-1041216" [ 897.818097] env[61964]: _type = "Task" [ 897.818097] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.833050] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.861380] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041214, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.027205] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041215, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.028695] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07830d8a-0e7f-46de-a149-8a2c40bebc69 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.033331] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a8df80-0e93-47c1-be6e-2ead12425575 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.052592] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd82cb97-998c-42d8-ac91-80304801243d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.056879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4ecb46-2251-4695-8b4e-6fb08e3dbdbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.063667] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.097336] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56166387-16af-4da8-bbc9-de717fe6a724 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.104911] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e27684-d081-4737-9dd6-52e787e49b45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.119773] env[61964]: DEBUG nova.compute.provider_tree [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.293671] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.327839] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.189296} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.328018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.328765] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4134d2-2aae-4e1b-aa30-e97975213a99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.349551] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.349551] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1baac177-a9c4-4c5b-aae0-3b6fbdd10b79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.372478] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041214, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.373895] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 898.373895] env[61964]: value = "task-1041217" [ 898.373895] env[61964]: _type = "Task" [ 898.373895] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.382660] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.525689] env[61964]: DEBUG oslo_vmware.api [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041215, 'name': PowerOnVM_Task, 'duration_secs': 0.652546} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.526013] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 898.526255] env[61964]: INFO nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Took 6.31 seconds to spawn the instance on the hypervisor. [ 898.526446] env[61964]: DEBUG nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 898.527233] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24ffa0c-b211-40e2-a4cb-3d514a48890f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.599488] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 898.599614] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da33e6b7-e104-4c58-a584-db2f60364493 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.605859] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 898.605859] env[61964]: value = "task-1041218" [ 898.605859] env[61964]: _type = "Task" [ 898.605859] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.613679] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041218, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.622728] env[61964]: DEBUG nova.scheduler.client.report [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.873908] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041214, 'name': CreateSnapshot_Task, 'duration_secs': 1.152509} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.874373] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 898.878032] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed0343d-fcdd-4ef9-ab24-a1ae73e24fed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.893512] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041217, 'name': ReconfigVM_Task, 'duration_secs': 0.26505} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.899133] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Reconfigured VM instance instance-0000004f to attach disk [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.900014] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0bf08db-3950-4f8f-be8b-c83764505e0a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.906985] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 898.906985] env[61964]: value = "task-1041219" [ 898.906985] env[61964]: _type = "Task" [ 898.906985] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.917232] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041219, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.044510] env[61964]: INFO nova.compute.manager [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Took 32.96 seconds to build instance. [ 899.125039] env[61964]: DEBUG oslo_vmware.api [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041218, 'name': PowerOnVM_Task, 'duration_secs': 0.403877} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.125387] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 899.125640] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f165044-1a03-4b46-b417-357b63fe7027 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance '84821045-27b2-4a99-87f9-988b8615b83f' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 899.134264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.134855] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 899.137967] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.244s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.140653] env[61964]: INFO nova.compute.claims [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.412134] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.412570] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8debc128-68d4-4d6b-b26c-e063211112fb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.423815] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041219, 'name': Rename_Task, 'duration_secs': 0.223092} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.425042] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.425380] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 899.425380] env[61964]: value = "task-1041220" [ 899.425380] env[61964]: _type = "Task" [ 899.425380] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.425605] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c09e9e5e-3763-4f89-b01d-88263719a9aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.435050] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041220, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.442936] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 899.442936] env[61964]: value = "task-1041221" [ 899.442936] env[61964]: _type = "Task" [ 899.442936] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.449886] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.547478] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d6c8f416-9f4f-4450-9e31-f6095bd2a9a6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.466s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.647766] env[61964]: DEBUG nova.compute.utils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.649536] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 899.649627] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 899.700097] env[61964]: DEBUG nova.policy [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a86108ab0b3b405bbdade80bea001c21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34421874d16248e89d850bfde2677d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.937426] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041220, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.952757] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041221, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.010804] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Successfully created port: 3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.152867] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 900.442540] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041220, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.452627] env[61964]: DEBUG oslo_vmware.api [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041221, 'name': PowerOnVM_Task, 'duration_secs': 0.628055} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.452892] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.453113] env[61964]: INFO nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Took 5.87 seconds to spawn the instance on the hypervisor. [ 900.453361] env[61964]: DEBUG nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 900.454072] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b603256-6253-4fb0-9f9f-7e13b4cd222b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.457777] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4203b5-f0f8-4907-8bad-8e32c3c1e084 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.468087] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eee07d-1b44-43ec-bbcb-f396ebf5e4cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.499766] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe86d6a-3c58-4f0e-8191-dcb75d32967b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.506865] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e397aa8-08fd-46c4-b81f-cade35761033 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.520266] env[61964]: DEBUG nova.compute.provider_tree [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.938582] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041220, 'name': CloneVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.979750] env[61964]: INFO nova.compute.manager [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Took 34.64 seconds to build instance. [ 901.024222] env[61964]: DEBUG nova.scheduler.client.report [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 901.165280] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 901.201754] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 901.202098] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.202325] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 901.202575] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.202777] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 901.202972] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 901.203220] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 901.203417] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 901.203605] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 901.203792] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 901.203972] env[61964]: DEBUG nova.virt.hardware [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 901.204910] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2ff524-2e48-4ba6-a7ea-79c5513c94b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.213889] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbed5db6-e6f6-487c-b474-1a85a29fed72 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.348191] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.349537] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.349797] env[61964]: DEBUG nova.compute.manager [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Going to confirm migration 3 {{(pid=61964) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5218}} [ 901.447762] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041220, 'name': CloneVM_Task, 'duration_secs': 1.613824} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.448137] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Created linked-clone VM from snapshot [ 901.449212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91515836-db3c-48b6-ad32-4cb2d1d11880 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.457994] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Uploading image 28db7f9c-4e35-4a8c-a942-aadd3221f474 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 901.482614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ba992594-4195-47c1-9b0f-7ccad0da3fe7 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.160s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.489787] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.489787] env[61964]: value = "vm-230518" [ 901.489787] env[61964]: _type = "VirtualMachine" [ 901.489787] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.490100] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-34dfa06f-4e49-4f84-a408-5f0cb5cb0f9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.496931] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lease: (returnval){ [ 901.496931] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e528bd-c0b4-d5a5-49aa-2ee5be0e744f" [ 901.496931] env[61964]: _type = "HttpNfcLease" [ 901.496931] env[61964]: } obtained for exporting VM: (result){ [ 901.496931] env[61964]: value = "vm-230518" [ 901.496931] env[61964]: _type = "VirtualMachine" [ 901.496931] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.497243] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the lease: (returnval){ [ 901.497243] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e528bd-c0b4-d5a5-49aa-2ee5be0e744f" [ 901.497243] env[61964]: _type = "HttpNfcLease" [ 901.497243] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.504698] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.504698] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e528bd-c0b4-d5a5-49aa-2ee5be0e744f" [ 901.504698] env[61964]: _type = "HttpNfcLease" [ 901.504698] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.529414] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.529958] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 901.532791] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.227s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.534349] env[61964]: INFO nova.compute.claims [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.751019] env[61964]: DEBUG nova.compute.manager [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Received event network-vif-plugged-3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 901.751019] env[61964]: DEBUG oslo_concurrency.lockutils [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] Acquiring lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.751019] env[61964]: DEBUG oslo_concurrency.lockutils [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.751019] env[61964]: DEBUG oslo_concurrency.lockutils [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.751019] env[61964]: DEBUG nova.compute.manager [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] No waiting events found dispatching network-vif-plugged-3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 901.751791] env[61964]: WARNING nova.compute.manager [req-279764e9-e842-4c47-a843-3b46bc0ce031 req-e0a6d214-3816-4104-842f-a26e9ea0d503 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Received unexpected event network-vif-plugged-3e133918-bc4f-4a52-ac44-586ec8a513ed for instance with vm_state building and task_state spawning. [ 901.862209] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Successfully updated port: 3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.901637] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.901637] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.901637] env[61964]: DEBUG nova.network.neutron [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.901637] env[61964]: DEBUG nova.objects.instance [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'info_cache' on Instance uuid 84821045-27b2-4a99-87f9-988b8615b83f {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.010182] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.010182] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e528bd-c0b4-d5a5-49aa-2ee5be0e744f" [ 902.010182] env[61964]: _type = "HttpNfcLease" [ 902.010182] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.010182] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.010182] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e528bd-c0b4-d5a5-49aa-2ee5be0e744f" [ 902.010182] env[61964]: _type = "HttpNfcLease" [ 902.010182] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.010182] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b3e5e9-ff95-4f39-abbe-d0773d1845ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.019766] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.019766] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.085504] env[61964]: DEBUG nova.compute.utils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.093023] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 902.093023] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 902.149085] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fd8c1589-5a34-4e8d-b8b8-147161fdd01f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.168525] env[61964]: DEBUG nova.policy [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.196920] env[61964]: INFO nova.compute.manager [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Rebuilding instance [ 902.248839] env[61964]: DEBUG nova.compute.manager [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.250684] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd1a580-ccf2-4717-9e48-1a69c5ce5954 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.369733] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.369884] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquired lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.370032] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.525465] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Successfully created port: 36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.593223] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 902.942562] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f1dc9f-3b2b-4df8-8cdd-957a0c122fb8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.952183] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebff4c24-0c36-4ac4-af3f-b24f3ecd3c03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.991647] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132491a5-8c65-4571-a94f-86af160c6dab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.999689] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f67b443-646a-409f-99c6-7a4ffb721767 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.019154] env[61964]: DEBUG nova.compute.provider_tree [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.111864] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.266668] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 903.266980] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-016156da-9a05-42d8-af79-931a05abdc24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.277651] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 903.277651] env[61964]: value = "task-1041223" [ 903.277651] env[61964]: _type = "Task" [ 903.277651] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.284756] env[61964]: DEBUG nova.network.neutron [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [{"id": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "address": "fa:16:3e:30:78:4d", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a884ac5-73", "ovs_interfaceid": "2a884ac5-7364-4168-bf0a-d947f84b8b92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.289895] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041223, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.427550] env[61964]: DEBUG nova.network.neutron [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updating instance_info_cache with network_info: [{"id": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "address": "fa:16:3e:bf:dd:3d", "network": {"id": "c2ed0c35-ba3a-4527-ba38-9a1914bf917a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-13377105-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34421874d16248e89d850bfde2677d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e133918-bc", "ovs_interfaceid": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.522990] env[61964]: DEBUG nova.scheduler.client.report [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.612965] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 903.646135] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 903.646342] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.646504] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 903.646742] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.646944] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 903.647151] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 903.647369] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 903.647540] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 903.647726] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 903.647905] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 903.648111] env[61964]: DEBUG nova.virt.hardware [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 903.649050] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6478f5c-68b2-4c7e-85e0-19fc9da85c65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.659027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec89cf3-c86d-4218-a93c-8f8b4b13d06c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.780040] env[61964]: DEBUG nova.compute.manager [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Received event network-changed-3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 903.780290] env[61964]: DEBUG nova.compute.manager [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Refreshing instance network info cache due to event network-changed-3e133918-bc4f-4a52-ac44-586ec8a513ed. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 903.780484] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] Acquiring lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.790811] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-84821045-27b2-4a99-87f9-988b8615b83f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.791193] env[61964]: DEBUG nova.objects.instance [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'migration_context' on Instance uuid 84821045-27b2-4a99-87f9-988b8615b83f {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.792961] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041223, 'name': PowerOffVM_Task, 'duration_secs': 0.1894} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.793120] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 903.793243] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 903.794130] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2087a2e-801f-4477-a188-23baa17b409b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.803017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 903.803017] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a16968e-fb48-4922-b2bd-37f24dc7dd36 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.829820] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 903.830078] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 903.830269] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleting the datastore file [datastore2] ba0f227f-640a-4765-a84e-61bd81128356 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.830585] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-919830a7-8af6-488c-95f3-09500b145b00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.838027] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 903.838027] env[61964]: value = "task-1041225" [ 903.838027] env[61964]: _type = "Task" [ 903.838027] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.846165] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041225, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.932636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Releasing lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.932985] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Instance network_info: |[{"id": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "address": "fa:16:3e:bf:dd:3d", "network": {"id": "c2ed0c35-ba3a-4527-ba38-9a1914bf917a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-13377105-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34421874d16248e89d850bfde2677d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e133918-bc", "ovs_interfaceid": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 903.933333] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] Acquired lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.933524] env[61964]: DEBUG nova.network.neutron [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Refreshing network info cache for port 3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 903.934834] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:dd:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e133918-bc4f-4a52-ac44-586ec8a513ed', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.943668] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Creating folder: Project (34421874d16248e89d850bfde2677d43). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 903.943668] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a9e5842-f353-45de-9b52-6c4421438b57 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.957066] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Created folder: Project (34421874d16248e89d850bfde2677d43) in parent group-v230360. [ 903.957355] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Creating folder: Instances. Parent ref: group-v230519. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 903.957889] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83168ee6-83ea-48db-affd-a950871bc210 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.968033] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Created folder: Instances in parent group-v230519. [ 903.968270] env[61964]: DEBUG oslo.service.loopingcall [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.968450] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 903.968700] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-494a0262-9b79-4044-9b84-7682cf8c5080 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.990741] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.990741] env[61964]: value = "task-1041228" [ 903.990741] env[61964]: _type = "Task" [ 903.990741] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.999153] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041228, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.028845] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.029436] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 904.032631] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.358s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.034205] env[61964]: INFO nova.compute.claims [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.132960] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.133264] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.294029] env[61964]: DEBUG nova.objects.base [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Object Instance<84821045-27b2-4a99-87f9-988b8615b83f> lazy-loaded attributes: info_cache,migration_context {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.297020] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b63154-57e0-4241-859c-feeececce5a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.315357] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e40d6066-3ae7-4288-ab13-4df3856fc8ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.321231] env[61964]: DEBUG oslo_vmware.api [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 904.321231] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293306e-14d9-e29b-37f1-b50acf0d110b" [ 904.321231] env[61964]: _type = "Task" [ 904.321231] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.329724] env[61964]: DEBUG oslo_vmware.api [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293306e-14d9-e29b-37f1-b50acf0d110b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.349503] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041225, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213256} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.349777] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.350634] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 904.351061] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 904.502951] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041228, 'name': CreateVM_Task, 'duration_secs': 0.359985} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.503144] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 904.503863] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.504072] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.504597] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.504677] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8391559-7596-42cf-9c64-d4d9ec0d2cdb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.509608] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 904.509608] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ca1de6-1c94-7c76-8b66-aee0b0f800dc" [ 904.509608] env[61964]: _type = "Task" [ 904.509608] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.517869] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ca1de6-1c94-7c76-8b66-aee0b0f800dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.538806] env[61964]: DEBUG nova.compute.utils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.540194] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 904.540385] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.594981] env[61964]: DEBUG nova.policy [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.636087] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 904.663857] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Successfully updated port: 36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.798684] env[61964]: DEBUG nova.network.neutron [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updated VIF entry in instance network info cache for port 3e133918-bc4f-4a52-ac44-586ec8a513ed. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 904.799113] env[61964]: DEBUG nova.network.neutron [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updating instance_info_cache with network_info: [{"id": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "address": "fa:16:3e:bf:dd:3d", "network": {"id": "c2ed0c35-ba3a-4527-ba38-9a1914bf917a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-13377105-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34421874d16248e89d850bfde2677d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e133918-bc", "ovs_interfaceid": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.832662] env[61964]: DEBUG oslo_vmware.api [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293306e-14d9-e29b-37f1-b50acf0d110b, 'name': SearchDatastore_Task, 'duration_secs': 0.008426} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.832844] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.027794] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ca1de6-1c94-7c76-8b66-aee0b0f800dc, 'name': SearchDatastore_Task, 'duration_secs': 0.027126} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.028134] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.028377] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.028621] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.028771] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.028956] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.029264] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22cbef10-b26a-4479-98f9-8795052fdcf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.038361] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.038544] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.039274] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be171194-908b-48c9-8f38-4977bfc51edf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.044540] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 905.044540] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3a08e-cc83-ac32-d61a-4a3f6749d643" [ 905.044540] env[61964]: _type = "Task" [ 905.044540] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.051823] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 905.056846] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3a08e-cc83-ac32-d61a-4a3f6749d643, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.157895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.166875] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.167063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.167250] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.208777] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Successfully created port: 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.303046] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b67a20d-c0b1-4abe-9149-e696f3e744d9 req-717b0e8d-17d6-4a19-b7b1-1adb0f8a27dd service nova] Releasing lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.321823] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df055f89-a736-45ec-81f4-681ce0e1c616 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.330422] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b83755-cc32-4018-9c10-a799c0f19bca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.365189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf1a4ee-0e34-4420-9f0a-b03bf5718860 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.374757] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84caabe-219b-40ab-b116-4b8185076076 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.393484] env[61964]: DEBUG nova.compute.provider_tree [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.409498] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 905.409990] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.409990] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.409990] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.409990] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.409990] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 905.410820] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 905.411556] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 905.411556] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 905.411556] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 905.411556] env[61964]: DEBUG nova.virt.hardware [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 905.412409] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283f7eec-034e-4297-876e-790d831e6b86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.424117] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6171d89a-ffb1-4005-9afa-d834bec0e590 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.445538] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance VIF info [] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.451695] env[61964]: DEBUG oslo.service.loopingcall [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.452043] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 905.452282] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7791281c-df6a-45e7-92a3-2f10bc1e00e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.471134] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.471134] env[61964]: value = "task-1041229" [ 905.471134] env[61964]: _type = "Task" [ 905.471134] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.479047] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041229, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.555643] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a3a08e-cc83-ac32-d61a-4a3f6749d643, 'name': SearchDatastore_Task, 'duration_secs': 0.031003} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.556585] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbe086cd-d76a-4938-aab5-1f9337da97dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.566704] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 905.566704] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dd6fcf-09e4-7094-ad11-948366af76f3" [ 905.566704] env[61964]: _type = "Task" [ 905.566704] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.579299] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dd6fcf-09e4-7094-ad11-948366af76f3, 'name': SearchDatastore_Task, 'duration_secs': 0.01197} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.579718] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.580132] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8/f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 905.580542] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c32c6846-6d9e-4978-a302-bd8319a770c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.587855] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 905.587855] env[61964]: value = "task-1041230" [ 905.587855] env[61964]: _type = "Task" [ 905.587855] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.596073] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.701789] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.811990] env[61964]: DEBUG nova.compute.manager [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Received event network-vif-plugged-36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 905.812276] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Acquiring lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.812503] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.812811] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.812917] env[61964]: DEBUG nova.compute.manager [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] No waiting events found dispatching network-vif-plugged-36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 905.813104] env[61964]: WARNING nova.compute.manager [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Received unexpected event network-vif-plugged-36f6ac10-084a-4796-9f7d-cc945a356b01 for instance with vm_state building and task_state spawning. [ 905.813272] env[61964]: DEBUG nova.compute.manager [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Received event network-changed-36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 905.813432] env[61964]: DEBUG nova.compute.manager [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Refreshing instance network info cache due to event network-changed-36f6ac10-084a-4796-9f7d-cc945a356b01. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 905.813616] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Acquiring lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.851523] env[61964]: DEBUG nova.network.neutron [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Updating instance_info_cache with network_info: [{"id": "36f6ac10-084a-4796-9f7d-cc945a356b01", "address": "fa:16:3e:32:b7:41", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36f6ac10-08", "ovs_interfaceid": "36f6ac10-084a-4796-9f7d-cc945a356b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.897911] env[61964]: DEBUG nova.scheduler.client.report [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.982321] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041229, 'name': CreateVM_Task, 'duration_secs': 0.32104} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.982508] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 905.983037] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.983177] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.983523] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.984195] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00302e2c-4f47-4d1c-a23b-5222aeae1b59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.990504] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 905.990504] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4bd6f-6e9a-7050-57d4-2bbb6cb2c615" [ 905.990504] env[61964]: _type = "Task" [ 905.990504] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.999199] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4bd6f-6e9a-7050-57d4-2bbb6cb2c615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.068292] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 906.097035] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 906.097035] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.097035] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 906.097035] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.097035] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 906.097282] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 906.097378] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 906.097551] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 906.097735] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 906.097888] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 906.099508] env[61964]: DEBUG nova.virt.hardware [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 906.099508] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092ec10b-6059-44a9-b896-28f059555779 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.110495] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c257fd73-fa5e-4eb3-9008-f87edad5f402 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.114966] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041230, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.354022] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.354394] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Instance network_info: |[{"id": "36f6ac10-084a-4796-9f7d-cc945a356b01", "address": "fa:16:3e:32:b7:41", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36f6ac10-08", "ovs_interfaceid": "36f6ac10-084a-4796-9f7d-cc945a356b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 906.354744] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Acquired lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.355018] env[61964]: DEBUG nova.network.neutron [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Refreshing network info cache for port 36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.356198] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:b7:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '36f6ac10-084a-4796-9f7d-cc945a356b01', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.363659] env[61964]: DEBUG oslo.service.loopingcall [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.364686] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 906.364933] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6382f2b-d68a-4f81-bf75-85b5a2e67b53 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.389220] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.389220] env[61964]: value = "task-1041231" [ 906.389220] env[61964]: _type = "Task" [ 906.389220] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.396913] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041231, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.403232] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.403773] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 906.406939] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.580s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.407229] env[61964]: DEBUG nova.objects.instance [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'resources' on Instance uuid 50b04cbf-9064-4010-b350-638e8096bb96 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.501354] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52e4bd6f-6e9a-7050-57d4-2bbb6cb2c615, 'name': SearchDatastore_Task, 'duration_secs': 0.055057} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.501681] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.501919] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.502235] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.502335] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.502516] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.502882] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7bc5e568-60ef-410e-858c-b997731d349d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.511598] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.511720] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 906.512459] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bd9f17f-2cd3-44c8-870e-805bf72b2fab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.517610] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 906.517610] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fa018e-5aea-a444-c1c2-47c092439c10" [ 906.517610] env[61964]: _type = "Task" [ 906.517610] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.525532] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fa018e-5aea-a444-c1c2-47c092439c10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.598507] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541796} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.598794] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8/f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 906.598997] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.599291] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-430a6b45-f02f-47d2-8945-4a066769c21a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.607252] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 906.607252] env[61964]: value = "task-1041232" [ 906.607252] env[61964]: _type = "Task" [ 906.607252] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.615535] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.698777] env[61964]: DEBUG nova.compute.manager [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-vif-plugged-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 906.698777] env[61964]: DEBUG oslo_concurrency.lockutils [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.698777] env[61964]: DEBUG oslo_concurrency.lockutils [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.698777] env[61964]: DEBUG oslo_concurrency.lockutils [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.698777] env[61964]: DEBUG nova.compute.manager [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] No waiting events found dispatching network-vif-plugged-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 906.698777] env[61964]: WARNING nova.compute.manager [req-1975f07d-5a7b-47b4-a5a1-a9e2e449d16d req-e137ab7f-7091-4577-aa95-099b793063e1 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received unexpected event network-vif-plugged-083961bb-ae89-4d60-8ea8-3d5454613407 for instance with vm_state building and task_state spawning. [ 906.796760] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Successfully updated port: 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.901270] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041231, 'name': CreateVM_Task, 'duration_secs': 0.350057} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.901449] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 906.902160] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.902341] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.902684] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.902951] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ff84bc7-d839-4728-9c93-397d296e2fe0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.907689] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 906.907689] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c5ce04-9edd-918a-ee6a-5202b41a22d1" [ 906.907689] env[61964]: _type = "Task" [ 906.907689] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.911822] env[61964]: DEBUG nova.compute.utils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.917931] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 906.917931] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 906.925154] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c5ce04-9edd-918a-ee6a-5202b41a22d1, 'name': SearchDatastore_Task, 'duration_secs': 0.011627} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.926155] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.926314] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.926553] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.958007] env[61964]: DEBUG nova.policy [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf85b27f46c94956982ca3c713aef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b79a033963e04e02a98df18f8429263f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.030661] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52fa018e-5aea-a444-c1c2-47c092439c10, 'name': SearchDatastore_Task, 'duration_secs': 0.013082} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.031663] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35c40e0e-8287-4ebe-b8cd-b927d626a8d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.040698] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 907.040698] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b738c3-399b-9491-599b-6926d082277a" [ 907.040698] env[61964]: _type = "Task" [ 907.040698] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.049822] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b738c3-399b-9491-599b-6926d082277a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.117701] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069235} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.118696] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.119259] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72089cba-78af-4d68-8258-989eeffdd940 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.151566] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8/f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.154872] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb2f6eec-466b-4ba1-a28b-534af4f4e787 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.177373] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 907.177373] env[61964]: value = "task-1041233" [ 907.177373] env[61964]: _type = "Task" [ 907.177373] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.185531] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.186449] env[61964]: DEBUG nova.network.neutron [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Updated VIF entry in instance network info cache for port 36f6ac10-084a-4796-9f7d-cc945a356b01. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 907.186794] env[61964]: DEBUG nova.network.neutron [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Updating instance_info_cache with network_info: [{"id": "36f6ac10-084a-4796-9f7d-cc945a356b01", "address": "fa:16:3e:32:b7:41", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap36f6ac10-08", "ovs_interfaceid": "36f6ac10-084a-4796-9f7d-cc945a356b01", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.278758] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8949e661-3ba9-46fe-953d-cacf78a0588b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.286834] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996bcdf8-4e56-4f49-b00c-0f266f786756 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.319689] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.319847] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.320093] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.322715] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce463ce4-90e7-4ff6-a95c-441bfd637692 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.330891] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df04611e-4093-4681-ae1b-8905571a75a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.346179] env[61964]: DEBUG nova.compute.provider_tree [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.405821] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Successfully created port: d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.418751] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 907.559172] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b738c3-399b-9491-599b-6926d082277a, 'name': SearchDatastore_Task, 'duration_secs': 0.015383} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.559172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.559172] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 907.559172] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.559172] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.559172] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdbeaedc-6774-496d-a1c5-1325ddc2c1d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.562264] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ada51dd6-9472-44c4-ba7e-900928db2486 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.570083] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 907.570083] env[61964]: value = "task-1041234" [ 907.570083] env[61964]: _type = "Task" [ 907.570083] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.574579] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.574813] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 907.575933] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7360bdf-6ffc-4954-8685-17b390f8a1a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.581596] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.586827] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 907.586827] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b9e790-e80a-aca9-861f-6afc839d6a35" [ 907.586827] env[61964]: _type = "Task" [ 907.586827] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.594328] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b9e790-e80a-aca9-861f-6afc839d6a35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.690345] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041233, 'name': ReconfigVM_Task, 'duration_secs': 0.288978} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.690757] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Reconfigured VM instance instance-00000050 to attach disk [datastore2] f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8/f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.691781] env[61964]: DEBUG oslo_concurrency.lockutils [req-6da6fc99-bc8f-4b54-a9fa-1910564dafa1 req-b7995d47-6900-4131-84e7-795475d8f9ec service nova] Releasing lock "refresh_cache-4d9c24cf-e11b-4d95-ad41-538d91279689" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.692227] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6b8c9da-026d-43ea-b7ff-1c0cbe1d666c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.698553] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 907.698553] env[61964]: value = "task-1041235" [ 907.698553] env[61964]: _type = "Task" [ 907.698553] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.706733] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041235, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.867772] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 907.871551] env[61964]: ERROR nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [req-4bc5636b-c656-4b14-82d9-98efe066f4ea] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4bc5636b-c656-4b14-82d9-98efe066f4ea"}]} [ 907.888725] env[61964]: DEBUG nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 907.906885] env[61964]: DEBUG nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 907.907142] env[61964]: DEBUG nova.compute.provider_tree [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.920559] env[61964]: DEBUG nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 907.940702] env[61964]: DEBUG nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 908.067737] env[61964]: DEBUG nova.network.neutron [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.082561] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041234, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.099967] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b9e790-e80a-aca9-861f-6afc839d6a35, 'name': SearchDatastore_Task, 'duration_secs': 0.036818} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.101198] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa82628-62aa-4ef6-a908-c753d69ffbe2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.107442] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 908.107442] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5217c258-9ad1-2c29-0d85-09a10ef2266c" [ 908.107442] env[61964]: _type = "Task" [ 908.107442] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.116569] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5217c258-9ad1-2c29-0d85-09a10ef2266c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.211557] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041235, 'name': Rename_Task, 'duration_secs': 0.153148} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.214304] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 908.215332] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d890bac-3221-4ac1-adee-2d9f7679e917 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.223985] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 908.223985] env[61964]: value = "task-1041236" [ 908.223985] env[61964]: _type = "Task" [ 908.223985] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.232571] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.234121] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0784dabb-9f49-4671-beb3-9a1ea80dbf24 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.241122] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c041560-a5c1-4fe6-9769-293ace404ee9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.274358] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8c641f-4525-4394-a988-4f53eaf30ca7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.282426] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f84cc7-fba0-437b-890e-4645dbfed00f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.296898] env[61964]: DEBUG nova.compute.provider_tree [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.429753] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 908.459841] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 908.460196] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.460375] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.460575] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.460828] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.461029] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 908.461249] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 908.461454] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 908.461579] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 908.461777] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 908.461958] env[61964]: DEBUG nova.virt.hardware [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 908.462896] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdf3d93-1d1e-411c-a640-e394fbb4dfc5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.470926] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d64b2a0-7caa-42f9-921b-754cd1ef2b2f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.570837] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.571247] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Instance network_info: |[{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 908.571818] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:3b:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '083961bb-ae89-4d60-8ea8-3d5454613407', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.581752] env[61964]: DEBUG oslo.service.loopingcall [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.584849] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 908.584849] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2006a910-326d-4b2c-892d-3fb0c291a466 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.606957] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553747} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.608298] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 908.608554] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.608833] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.608833] env[61964]: value = "task-1041237" [ 908.608833] env[61964]: _type = "Task" [ 908.608833] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.609052] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12612b7b-cc3a-4a05-8517-52e527f51496 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.622200] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5217c258-9ad1-2c29-0d85-09a10ef2266c, 'name': SearchDatastore_Task, 'duration_secs': 0.028892} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.626616] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.626935] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4d9c24cf-e11b-4d95-ad41-538d91279689/4d9c24cf-e11b-4d95-ad41-538d91279689.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 908.627362] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 908.627362] env[61964]: value = "task-1041238" [ 908.627362] env[61964]: _type = "Task" [ 908.627362] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.627656] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041237, 'name': CreateVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.627836] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a706381-72dd-473d-a95e-98f1b9d134ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.638535] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.640245] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 908.640245] env[61964]: value = "task-1041239" [ 908.640245] env[61964]: _type = "Task" [ 908.640245] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.651696] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.731795] env[61964]: DEBUG nova.compute.manager [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 908.732014] env[61964]: DEBUG nova.compute.manager [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 908.732291] env[61964]: DEBUG oslo_concurrency.lockutils [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.732430] env[61964]: DEBUG oslo_concurrency.lockutils [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.732599] env[61964]: DEBUG nova.network.neutron [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 908.740618] env[61964]: DEBUG oslo_vmware.api [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041236, 'name': PowerOnVM_Task, 'duration_secs': 0.476967} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.741523] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.741754] env[61964]: INFO nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Took 7.58 seconds to spawn the instance on the hypervisor. [ 908.741964] env[61964]: DEBUG nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.742841] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0caf9ad6-2c97-4ca6-99fa-4af63ef64d48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.799645] env[61964]: DEBUG nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.868845] env[61964]: DEBUG nova.compute.manager [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Received event network-vif-plugged-d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 908.869729] env[61964]: DEBUG oslo_concurrency.lockutils [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.870088] env[61964]: DEBUG oslo_concurrency.lockutils [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.870322] env[61964]: DEBUG oslo_concurrency.lockutils [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.870678] env[61964]: DEBUG nova.compute.manager [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] No waiting events found dispatching network-vif-plugged-d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 908.870810] env[61964]: WARNING nova.compute.manager [req-d366b509-ba56-4b8d-9d5a-6666deca15a8 req-7e60562b-575b-4894-a8d9-d45082019b8c service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Received unexpected event network-vif-plugged-d9618c13-facd-40f7-b201-776d03dffe57 for instance with vm_state building and task_state spawning. [ 908.984436] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Successfully updated port: d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.123899] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041237, 'name': CreateVM_Task, 'duration_secs': 0.399052} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.124152] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 909.125176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.125176] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.125565] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 909.125888] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e42421f-50ac-434f-8e7d-ff97bed894df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.131450] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 909.131450] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f169db-e7e5-7385-59e3-2943a5d1a2e6" [ 909.131450] env[61964]: _type = "Task" [ 909.131450] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.143101] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094044} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.151602] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.151989] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f169db-e7e5-7385-59e3-2943a5d1a2e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.152765] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f623de1-b296-4af5-9eef-074f860f2587 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.160529] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041239, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.175939] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.176246] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72463ca6-62aa-43a1-a983-5257b09f40ac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.196673] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 909.196673] env[61964]: value = "task-1041240" [ 909.196673] env[61964]: _type = "Task" [ 909.196673] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.204765] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041240, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.263411] env[61964]: INFO nova.compute.manager [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Took 28.55 seconds to build instance. [ 909.305906] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.899s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.308276] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 24.144s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.327963] env[61964]: INFO nova.scheduler.client.report [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted allocations for instance 50b04cbf-9064-4010-b350-638e8096bb96 [ 909.487114] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.487409] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.487454] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.567199] env[61964]: DEBUG nova.network.neutron [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 909.567199] env[61964]: DEBUG nova.network.neutron [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.646048] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f169db-e7e5-7385-59e3-2943a5d1a2e6, 'name': SearchDatastore_Task, 'duration_secs': 0.027786} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.649430] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.650030] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.650030] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.650168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.650392] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.651138] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a357ad17-1dc1-4880-8bda-c785caf5a8b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.658279] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535346} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.658542] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 4d9c24cf-e11b-4d95-ad41-538d91279689/4d9c24cf-e11b-4d95-ad41-538d91279689.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 909.658778] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.660122] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fc082da-0e00-44c0-85af-d5f0be706545 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.662085] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.662268] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 909.663018] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-892091ba-a624-48fb-9a95-58372b57dd2d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.668529] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 909.668529] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5259d0ba-2681-bce3-e027-a82ee5a435a2" [ 909.668529] env[61964]: _type = "Task" [ 909.668529] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.673124] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 909.673124] env[61964]: value = "task-1041241" [ 909.673124] env[61964]: _type = "Task" [ 909.673124] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.679406] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5259d0ba-2681-bce3-e027-a82ee5a435a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.685271] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.706758] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041240, 'name': ReconfigVM_Task, 'duration_secs': 0.285315} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.707162] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Reconfigured VM instance instance-0000004f to attach disk [datastore2] ba0f227f-640a-4765-a84e-61bd81128356/ba0f227f-640a-4765-a84e-61bd81128356.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.707961] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52ba125c-1154-4d7b-ab69-6332015df647 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.715933] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 909.715933] env[61964]: value = "task-1041242" [ 909.715933] env[61964]: _type = "Task" [ 909.715933] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.724548] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041242, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.765870] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ed05b456-58fe-4ee9-9554-eece4fb55aeb tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.066s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.835225] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8dc44c8b-d1a3-489f-acfc-b758accab951 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "50b04cbf-9064-4010-b350-638e8096bb96" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.600s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.036107] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.069867] env[61964]: DEBUG oslo_concurrency.lockutils [req-c2168c1b-da05-489d-bc10-3d591989d7a9 req-c0430e52-dbde-4caf-9ef2-e33c35f46ddd service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.119570] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af39e541-fb7c-4db5-9ee1-102d6941506a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.132995] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d983259-a070-466a-9e54-471efe932ef7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.162439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ae5006-5e3f-4f13-b3e1-75e10288a062 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.180062] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d401132-62d7-4bd1-92da-7c1ada05f023 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.188532] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5259d0ba-2681-bce3-e027-a82ee5a435a2, 'name': SearchDatastore_Task, 'duration_secs': 0.012651} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.189246] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acb22526-e3f4-4fb6-969f-1c071aaf11bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.203207] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072116} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.204019] env[61964]: DEBUG nova.compute.provider_tree [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.205816] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.206639] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a05e865-0b40-4c1a-b011-a250fa13dcd7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.210491] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 910.210491] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ea8c00-a972-163d-db58-ce3b28f55e6c" [ 910.210491] env[61964]: _type = "Task" [ 910.210491] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.234262] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 4d9c24cf-e11b-4d95-ad41-538d91279689/4d9c24cf-e11b-4d95-ad41-538d91279689.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.238812] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1413e065-ddce-4994-8eb9-4b78d30c5385 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.262181] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ea8c00-a972-163d-db58-ce3b28f55e6c, 'name': SearchDatastore_Task, 'duration_secs': 0.0118} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.262181] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.262181] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 47378856-4cbf-4cf6-aecd-ae935885df3d/47378856-4cbf-4cf6-aecd-ae935885df3d.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 910.262181] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1032ce30-e310-428a-8412-51b92aab7052 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.267455] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041242, 'name': Rename_Task, 'duration_secs': 0.139307} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.269034] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.269375] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 910.269375] env[61964]: value = "task-1041243" [ 910.269375] env[61964]: _type = "Task" [ 910.269375] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.269643] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab74722b-1914-413e-91e6-44b7315f9321 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.277014] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 910.277014] env[61964]: value = "task-1041244" [ 910.277014] env[61964]: _type = "Task" [ 910.277014] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.283158] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 910.283158] env[61964]: value = "task-1041245" [ 910.283158] env[61964]: _type = "Task" [ 910.283158] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.283420] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.290351] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041244, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.293844] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.322237] env[61964]: DEBUG nova.network.neutron [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating instance_info_cache with network_info: [{"id": "d9618c13-facd-40f7-b201-776d03dffe57", "address": "fa:16:3e:5c:78:03", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9618c13-fa", "ovs_interfaceid": "d9618c13-facd-40f7-b201-776d03dffe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.427281] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.427704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.428039] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.428340] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.428632] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.431576] env[61964]: INFO nova.compute.manager [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Terminating instance [ 910.735640] env[61964]: ERROR nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [req-7810e8cc-9cc3-406d-bc60-fd31c1103b04] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7810e8cc-9cc3-406d-bc60-fd31c1103b04"}]} [ 910.757170] env[61964]: DEBUG nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 910.778468] env[61964]: DEBUG nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 910.778818] env[61964]: DEBUG nova.compute.provider_tree [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.795491] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041243, 'name': ReconfigVM_Task, 'duration_secs': 0.504664} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.801022] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 4d9c24cf-e11b-4d95-ad41-538d91279689/4d9c24cf-e11b-4d95-ad41-538d91279689.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.801022] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041244, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.801022] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df67dc6f-486d-4953-a6ac-cb157d68f3ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.807406] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041245, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.810214] env[61964]: DEBUG nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 910.814329] env[61964]: DEBUG nova.compute.manager [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Received event network-changed-3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 910.815803] env[61964]: DEBUG nova.compute.manager [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Refreshing instance network info cache due to event network-changed-3e133918-bc4f-4a52-ac44-586ec8a513ed. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 910.816164] env[61964]: DEBUG oslo_concurrency.lockutils [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] Acquiring lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.816391] env[61964]: DEBUG oslo_concurrency.lockutils [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] Acquired lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.816624] env[61964]: DEBUG nova.network.neutron [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Refreshing network info cache for port 3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.819963] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 910.819963] env[61964]: value = "task-1041246" [ 910.819963] env[61964]: _type = "Task" [ 910.819963] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.825021] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.825376] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Instance network_info: |[{"id": "d9618c13-facd-40f7-b201-776d03dffe57", "address": "fa:16:3e:5c:78:03", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9618c13-fa", "ovs_interfaceid": "d9618c13-facd-40f7-b201-776d03dffe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 910.826505] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:78:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97b68ed7-8461-4345-b064-96a1dde53a86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9618c13-facd-40f7-b201-776d03dffe57', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.834396] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating folder: Project (b79a033963e04e02a98df18f8429263f). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 910.836279] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e23abe5-b2d5-434d-bfbd-0a2a0ee20fb6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.842887] env[61964]: DEBUG nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 910.845688] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041246, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.856971] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created folder: Project (b79a033963e04e02a98df18f8429263f) in parent group-v230360. [ 910.856971] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating folder: Instances. Parent ref: group-v230525. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 910.856971] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af5c4b2e-2ad4-4d0c-b839-c93dc82dc0a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.867620] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created folder: Instances in parent group-v230525. [ 910.867929] env[61964]: DEBUG oslo.service.loopingcall [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.869029] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 910.869029] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5fa5d5c9-d885-4295-9fb0-92d5b338bc79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.894777] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.894777] env[61964]: value = "task-1041249" [ 910.894777] env[61964]: _type = "Task" [ 910.894777] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.903652] env[61964]: DEBUG nova.compute.manager [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Received event network-changed-d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 910.903846] env[61964]: DEBUG nova.compute.manager [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Refreshing instance network info cache due to event network-changed-d9618c13-facd-40f7-b201-776d03dffe57. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 910.904206] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] Acquiring lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.904370] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] Acquired lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.904540] env[61964]: DEBUG nova.network.neutron [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Refreshing network info cache for port d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.913932] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041249, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.937770] env[61964]: DEBUG nova.compute.manager [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 910.938052] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 910.938986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0a9ddd-e816-4414-bae7-c07155f1e51c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.950727] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 910.951046] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14da1b39-542b-4cd6-a585-f47f922dabea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.960266] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 910.960266] env[61964]: value = "task-1041250" [ 910.960266] env[61964]: _type = "Task" [ 910.960266] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.971803] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.062759] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "58f1ad74-ea72-45c8-b98c-c85674989342" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.063155] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.158411] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831970c5-1687-432c-a8e6-4a94ab563693 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.167747] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556ce85b-f655-4b5f-bce0-67d0cbfa6789 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.200849] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b96041-1f32-43f0-b821-ee13dc410423 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.209705] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35dc70a-2528-4412-84c3-60f781b39ba4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.227074] env[61964]: DEBUG nova.compute.provider_tree [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.286238] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041244, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534175} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.286726] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 47378856-4cbf-4cf6-aecd-ae935885df3d/47378856-4cbf-4cf6-aecd-ae935885df3d.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.286726] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.289587] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68f7a484-e709-48ee-b965-cda00bbcbf28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.297916] env[61964]: DEBUG oslo_vmware.api [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041245, 'name': PowerOnVM_Task, 'duration_secs': 0.643738} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.301209] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 911.301360] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.301686] env[61964]: DEBUG nova.compute.manager [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 911.302199] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 911.302199] env[61964]: value = "task-1041251" [ 911.302199] env[61964]: _type = "Task" [ 911.302199] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.303381] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13e2b5c-30d6-4c64-9632-632add4363f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.307410] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58236952-ba60-4d4f-bf2a-60363f47bd14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.318367] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 911.318539] env[61964]: ERROR oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk due to incomplete transfer. [ 911.322536] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f6b934d8-2292-478c-b31f-de3e7db2fb7e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.325929] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041251, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.337333] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041246, 'name': Rename_Task, 'duration_secs': 0.215994} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.338425] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 911.338705] env[61964]: DEBUG oslo_vmware.rw_handles [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52dcba23-9be7-4ff5-d760-6c83b3f11191/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 911.338890] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Uploaded image 28db7f9c-4e35-4a8c-a942-aadd3221f474 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 911.341449] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 911.341707] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75ba0110-74b2-475d-b066-a5df89041d63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.343258] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-add8a8e1-488e-4034-8f0f-499333c92b33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.349107] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 911.349107] env[61964]: value = "task-1041252" [ 911.349107] env[61964]: _type = "Task" [ 911.349107] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.350370] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 911.350370] env[61964]: value = "task-1041253" [ 911.350370] env[61964]: _type = "Task" [ 911.350370] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.361239] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041253, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.364339] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.404891] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041249, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.470588] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041250, 'name': PowerOffVM_Task, 'duration_secs': 0.315673} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.471094] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 911.471308] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 911.471584] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d3d0cad-153b-4b7a-8b7d-5089ecd0cb9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.565382] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 911.613054] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.613241] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.613437] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore1] 8bf586a2-0b56-4f5a-9f33-d537fba00d32 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.613748] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef067781-8f92-4c07-99bd-c9cef58daf08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.621181] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 911.621181] env[61964]: value = "task-1041255" [ 911.621181] env[61964]: _type = "Task" [ 911.621181] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.631245] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.635833] env[61964]: DEBUG nova.network.neutron [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updated VIF entry in instance network info cache for port 3e133918-bc4f-4a52-ac44-586ec8a513ed. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.636307] env[61964]: DEBUG nova.network.neutron [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updating instance_info_cache with network_info: [{"id": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "address": "fa:16:3e:bf:dd:3d", "network": {"id": "c2ed0c35-ba3a-4527-ba38-9a1914bf917a", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-13377105-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34421874d16248e89d850bfde2677d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e133918-bc", "ovs_interfaceid": "3e133918-bc4f-4a52-ac44-586ec8a513ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.645017] env[61964]: DEBUG nova.network.neutron [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updated VIF entry in instance network info cache for port d9618c13-facd-40f7-b201-776d03dffe57. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.645399] env[61964]: DEBUG nova.network.neutron [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating instance_info_cache with network_info: [{"id": "d9618c13-facd-40f7-b201-776d03dffe57", "address": "fa:16:3e:5c:78:03", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9618c13-fa", "ovs_interfaceid": "d9618c13-facd-40f7-b201-776d03dffe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.765515] env[61964]: DEBUG nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 109 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 911.765998] env[61964]: DEBUG nova.compute.provider_tree [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 109 to 110 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 911.766964] env[61964]: DEBUG nova.compute.provider_tree [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.820416] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041251, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080283} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.820802] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.821878] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f268024d-4c7e-4a97-a2cc-c8092860c9c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.850949] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 47378856-4cbf-4cf6-aecd-ae935885df3d/47378856-4cbf-4cf6-aecd-ae935885df3d.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.853696] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.854352] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2abc2e14-a25d-4e6e-91da-734848f14b61 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.880661] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041253, 'name': Destroy_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.885059] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.885401] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 911.885401] env[61964]: value = "task-1041256" [ 911.885401] env[61964]: _type = "Task" [ 911.885401] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.893818] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041256, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.905593] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041249, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.087833] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.138870] env[61964]: DEBUG oslo_vmware.api [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281803} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.139203] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 912.139420] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 912.139632] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 912.139819] env[61964]: INFO nova.compute.manager [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Took 1.20 seconds to destroy the instance on the hypervisor. [ 912.140096] env[61964]: DEBUG oslo.service.loopingcall [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.140323] env[61964]: DEBUG nova.compute.manager [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 912.140420] env[61964]: DEBUG nova.network.neutron [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 912.142458] env[61964]: DEBUG oslo_concurrency.lockutils [req-e7fa88bd-f34a-4ce9-a5e5-c25381cf16ed req-ecbe8fb6-7fca-4786-9e2f-0313e0076c16 service nova] Releasing lock "refresh_cache-f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.149567] env[61964]: DEBUG oslo_concurrency.lockutils [req-0f0e0352-6f60-4f72-b84c-8c56738a4075 req-f349699e-5e39-402a-aa1b-5346d7fb6e2f service nova] Releasing lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.365368] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.379487] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041253, 'name': Destroy_Task, 'duration_secs': 0.575202} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.379771] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Destroyed the VM [ 912.380067] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 912.380335] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e8e749cd-4c6f-4b0b-b695-400ab6bbe22b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.387781] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 912.387781] env[61964]: value = "task-1041257" [ 912.387781] env[61964]: _type = "Task" [ 912.387781] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.406778] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041257, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.407045] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041256, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.411871] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041249, 'name': CreateVM_Task, 'duration_secs': 1.472934} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.412048] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.412984] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.412984] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.413262] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.413527] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02bce756-01b4-44fe-bd3a-4621472b1666 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.421188] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 912.421188] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5203701b-2d37-0a97-125b-353e781d95ba" [ 912.421188] env[61964]: _type = "Task" [ 912.421188] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.431244] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5203701b-2d37-0a97-125b-353e781d95ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.771938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "ba0f227f-640a-4765-a84e-61bd81128356" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.771938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.772314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "ba0f227f-640a-4765-a84e-61bd81128356-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.772372] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.772589] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.775081] env[61964]: INFO nova.compute.manager [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Terminating instance [ 912.781258] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.473s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.781372] env[61964]: DEBUG nova.compute.manager [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61964) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5344}} [ 912.784795] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 26.740s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.863207] env[61964]: DEBUG nova.compute.manager [req-8532b5bb-1dce-43c4-a0ba-e33946ef8ed6 req-7e270fd2-fcfd-4c3f-8f02-06a26ec6f0a2 service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Received event network-vif-deleted-e2ef8b26-3c97-46b6-b6d2-407abf5737c9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 912.863207] env[61964]: INFO nova.compute.manager [req-8532b5bb-1dce-43c4-a0ba-e33946ef8ed6 req-7e270fd2-fcfd-4c3f-8f02-06a26ec6f0a2 service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Neutron deleted interface e2ef8b26-3c97-46b6-b6d2-407abf5737c9; detaching it from the instance and deleting it from the info cache [ 912.863207] env[61964]: DEBUG nova.network.neutron [req-8532b5bb-1dce-43c4-a0ba-e33946ef8ed6 req-7e270fd2-fcfd-4c3f-8f02-06a26ec6f0a2 service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.867592] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041252, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.898430] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041256, 'name': ReconfigVM_Task, 'duration_secs': 0.683504} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.899067] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 47378856-4cbf-4cf6-aecd-ae935885df3d/47378856-4cbf-4cf6-aecd-ae935885df3d.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.899979] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b560364e-4973-448a-b114-88d7d023f11d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.904327] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041257, 'name': RemoveSnapshot_Task, 'duration_secs': 0.489225} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.904913] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 912.905561] env[61964]: DEBUG nova.compute.manager [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 912.905970] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80df4934-21bf-4647-aa72-cf7f6e2e27b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.909864] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 912.909864] env[61964]: value = "task-1041258" [ 912.909864] env[61964]: _type = "Task" [ 912.909864] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.924608] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041258, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.925275] env[61964]: DEBUG nova.network.neutron [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.929963] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5203701b-2d37-0a97-125b-353e781d95ba, 'name': SearchDatastore_Task, 'duration_secs': 0.00938} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.930465] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.930701] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.930981] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.931168] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.931353] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.931617] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dce9ac8a-9caa-454b-acfb-c70105a68647 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.943275] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.943553] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 912.944359] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf25e140-df6f-4471-a554-43e69bc9deb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.950257] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 912.950257] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ccf347-0e9c-a638-c87e-3ba11bb2d2fc" [ 912.950257] env[61964]: _type = "Task" [ 912.950257] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.958925] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ccf347-0e9c-a638-c87e-3ba11bb2d2fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.279250] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "refresh_cache-ba0f227f-640a-4765-a84e-61bd81128356" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.279478] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "refresh_cache-ba0f227f-640a-4765-a84e-61bd81128356" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.279672] env[61964]: DEBUG nova.network.neutron [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.361347] env[61964]: INFO nova.scheduler.client.report [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted allocation for migration 8eb2d8ea-8ce7-4c30-9e2e-6127519d71b3 [ 913.368727] env[61964]: DEBUG oslo_vmware.api [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041252, 'name': PowerOnVM_Task, 'duration_secs': 1.517719} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.369192] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29481664-21e3-4cff-8f40-3138574a75c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.371358] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 913.371588] env[61964]: INFO nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Took 9.76 seconds to spawn the instance on the hypervisor. [ 913.371775] env[61964]: DEBUG nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.372579] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a03c92-73e3-4388-9e7a-03e2331fe00d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.388222] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8b139e-a732-492f-ba5c-ca844bea3724 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.424530] env[61964]: DEBUG nova.compute.manager [req-8532b5bb-1dce-43c4-a0ba-e33946ef8ed6 req-7e270fd2-fcfd-4c3f-8f02-06a26ec6f0a2 service nova] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Detach interface failed, port_id=e2ef8b26-3c97-46b6-b6d2-407abf5737c9, reason: Instance 8bf586a2-0b56-4f5a-9f33-d537fba00d32 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 913.425639] env[61964]: INFO nova.compute.manager [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Shelve offloading [ 913.430957] env[61964]: INFO nova.compute.manager [-] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Took 1.29 seconds to deallocate network for instance. [ 913.436567] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041258, 'name': Rename_Task, 'duration_secs': 0.300876} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.439139] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.440058] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10ab28ed-d98a-47e2-aa6d-c2784d39cb83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.446523] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 913.446523] env[61964]: value = "task-1041259" [ 913.446523] env[61964]: _type = "Task" [ 913.446523] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.457146] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.463908] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ccf347-0e9c-a638-c87e-3ba11bb2d2fc, 'name': SearchDatastore_Task, 'duration_secs': 0.008003} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.464604] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eed3ea73-ac93-4569-bf72-dc3cdb629366 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.469514] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 913.469514] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cc08a-2cff-7892-f2da-d2e95f45e163" [ 913.469514] env[61964]: _type = "Task" [ 913.469514] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.478005] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cc08a-2cff-7892-f2da-d2e95f45e163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.798800] env[61964]: DEBUG nova.network.neutron [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.803842] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Applying migration context for instance 84821045-27b2-4a99-87f9-988b8615b83f as it has an incoming, in-progress migration 8b2cfb5a-d6e6-49e9-8d5f-74cdad8f363a. Migration status is confirming {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 913.805686] env[61964]: INFO nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating resource usage from migration 8b2cfb5a-d6e6-49e9-8d5f-74cdad8f363a [ 913.824174] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance aef4c3a7-641a-4356-9187-ae4c082ccde9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.824360] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 8bf586a2-0b56-4f5a-9f33-d537fba00d32 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.824503] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.824570] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance df0a9a69-bd44-4da4-ba3a-9ba241c010a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.826188] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance f73b99bf-6a2b-4f21-b855-74965c95ed76 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.826188] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Migration 8b2cfb5a-d6e6-49e9-8d5f-74cdad8f363a is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 913.826188] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 84821045-27b2-4a99-87f9-988b8615b83f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.826188] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 52f358e4-beb2-4b96-8f96-5cd4da103eda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.826188] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance ba0f227f-640a-4765-a84e-61bd81128356 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 913.850150] env[61964]: DEBUG nova.network.neutron [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.870989] env[61964]: DEBUG oslo_concurrency.lockutils [None req-833d25df-b10d-4afd-a114-f525d30d71d8 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 32.136s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.893959] env[61964]: INFO nova.compute.manager [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Took 32.02 seconds to build instance. [ 913.905912] env[61964]: DEBUG nova.objects.instance [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.932196] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.932395] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b35203ab-1ff7-4611-8494-caad5e8b0a79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.940256] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 913.940256] env[61964]: value = "task-1041260" [ 913.940256] env[61964]: _type = "Task" [ 913.940256] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.941167] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.949872] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041260, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.958475] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041259, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.979738] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528cc08a-2cff-7892-f2da-d2e95f45e163, 'name': SearchDatastore_Task, 'duration_secs': 0.008536} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.980040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.980301] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 8a14232f-5cc6-4856-9291-eb0a1bd2ab69/8a14232f-5cc6-4856-9291-eb0a1bd2ab69.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 913.980734] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c8359f5-b282-41f6-83bc-f041fdaf62d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.987671] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 913.987671] env[61964]: value = "task-1041261" [ 913.987671] env[61964]: _type = "Task" [ 913.987671] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.997464] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041261, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.328143] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 63793ce6-d511-403e-8a4b-cad8c4157449 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 914.328386] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.328461] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 4d9c24cf-e11b-4d95-ad41-538d91279689 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.328543] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 47378856-4cbf-4cf6-aecd-ae935885df3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.328654] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 914.353259] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "refresh_cache-ba0f227f-640a-4765-a84e-61bd81128356" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.353640] env[61964]: DEBUG nova.compute.manager [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 914.353844] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.354820] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe483a4e-10f9-4f10-a3c3-5af3bc369127 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.364258] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 914.364553] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01cab35c-93c6-492a-a210-17cfba56de4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.372132] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 914.372132] env[61964]: value = "task-1041262" [ 914.372132] env[61964]: _type = "Task" [ 914.372132] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.382042] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.396874] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c08aad91-74be-4caf-84f9-edb585b03a75 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.530s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.413427] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.413636] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.413837] env[61964]: DEBUG nova.network.neutron [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.414073] env[61964]: DEBUG nova.objects.instance [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'info_cache' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.457899] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 914.458593] env[61964]: DEBUG nova.compute.manager [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.459087] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667b5d7e-fd52-433a-8514-d3f940d44b83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.466935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.467230] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.467518] env[61964]: DEBUG nova.network.neutron [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.472739] env[61964]: DEBUG oslo_vmware.api [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041259, 'name': PowerOnVM_Task, 'duration_secs': 0.648928} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.473016] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.473161] env[61964]: INFO nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Took 8.40 seconds to spawn the instance on the hypervisor. [ 914.473366] env[61964]: DEBUG nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.474303] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e582b43f-5c97-4f71-a77e-c3a199f94b62 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.496870] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041261, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488231} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.497174] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 8a14232f-5cc6-4856-9291-eb0a1bd2ab69/8a14232f-5cc6-4856-9291-eb0a1bd2ab69.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 914.497495] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.497671] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bda95385-1ce8-4590-b9f8-bfdb5368e830 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.506477] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 914.506477] env[61964]: value = "task-1041263" [ 914.506477] env[61964]: _type = "Task" [ 914.506477] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.515684] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.688654] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "2c77a07d-c949-4aef-8d6f-132969d58cee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.688898] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.831426] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance f5acabe4-674b-4c4c-85b4-b9270926017b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 914.882925] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041262, 'name': PowerOffVM_Task, 'duration_secs': 0.331301} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.883206] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.883392] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.883686] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e331986-f004-445e-9387-3631464ac367 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.911774] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.911990] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.912204] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleting the datastore file [datastore2] ba0f227f-640a-4765-a84e-61bd81128356 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.912466] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-535fc64e-aa67-4b8d-bbc1-96918d663966 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.917252] env[61964]: DEBUG nova.objects.base [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 914.920650] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 914.920650] env[61964]: value = "task-1041265" [ 914.920650] env[61964]: _type = "Task" [ 914.920650] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.929809] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.993993] env[61964]: INFO nova.compute.manager [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Took 32.71 seconds to build instance. [ 915.018328] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085705} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.018599] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.019390] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa500994-1a79-4a0d-878c-7474056b5124 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.043740] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 8a14232f-5cc6-4856-9291-eb0a1bd2ab69/8a14232f-5cc6-4856-9291-eb0a1bd2ab69.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.044384] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-342d677a-c73d-4c8a-a5e0-63c743b5fd3b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.064504] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 915.064504] env[61964]: value = "task-1041266" [ 915.064504] env[61964]: _type = "Task" [ 915.064504] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.073333] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.192089] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 915.262937] env[61964]: DEBUG nova.network.neutron [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updating instance_info_cache with network_info: [{"id": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "address": "fa:16:3e:b8:3b:74", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e14aef8-cd", "ovs_interfaceid": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.334809] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 003a5b74-2b8d-4e14-a7ee-db8006f81dfa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 915.435075] env[61964]: DEBUG oslo_vmware.api [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110347} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.435075] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.435075] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.435075] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.435075] env[61964]: INFO nova.compute.manager [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Took 1.08 seconds to destroy the instance on the hypervisor. [ 915.435075] env[61964]: DEBUG oslo.service.loopingcall [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.435075] env[61964]: DEBUG nova.compute.manager [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 915.435075] env[61964]: DEBUG nova.network.neutron [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.453701] env[61964]: DEBUG nova.network.neutron [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.496623] env[61964]: DEBUG oslo_concurrency.lockutils [None req-eda1edec-06b7-41ad-b998-e83002ef3ddf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.225s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.582675] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.699415] env[61964]: DEBUG nova.network.neutron [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [{"id": "8b2c44a6-df10-43fa-9afa-85167edbb650", "address": "fa:16:3e:f6:a0:e1", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2c44a6-df", "ovs_interfaceid": "8b2c44a6-df10-43fa-9afa-85167edbb650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.726464] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.766151] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.837903] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance d8f24177-b695-4d40-958d-9272b4f683cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 915.853594] env[61964]: DEBUG nova.compute.manager [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 915.853810] env[61964]: DEBUG nova.compute.manager [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 915.854041] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.854195] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.854359] env[61964]: DEBUG nova.network.neutron [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.956931] env[61964]: DEBUG nova.network.neutron [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.075913] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041266, 'name': ReconfigVM_Task, 'duration_secs': 0.952007} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.076266] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 8a14232f-5cc6-4856-9291-eb0a1bd2ab69/8a14232f-5cc6-4856-9291-eb0a1bd2ab69.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.076951] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bf9966f-689c-474e-b513-ab03a0e7cec2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.084086] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 916.084086] env[61964]: value = "task-1041267" [ 916.084086] env[61964]: _type = "Task" [ 916.084086] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.092661] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041267, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.139779] env[61964]: DEBUG nova.compute.manager [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received event network-vif-unplugged-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 916.140049] env[61964]: DEBUG oslo_concurrency.lockutils [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.140270] env[61964]: DEBUG oslo_concurrency.lockutils [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.140458] env[61964]: DEBUG oslo_concurrency.lockutils [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.140646] env[61964]: DEBUG nova.compute.manager [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] No waiting events found dispatching network-vif-unplugged-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 916.140831] env[61964]: WARNING nova.compute.manager [req-120b4c60-56a2-4489-99e5-d206545bc55d req-b0f1792d-2ca5-497f-8f3f-0664a3af792e service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received unexpected event network-vif-unplugged-9e14aef8-cd53-45ad-9b3c-152397bc01da for instance with vm_state shelved and task_state shelving_offloading. [ 916.184790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.186111] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcfcfa0-1fcf-40c7-8ca3-d478d303d110 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.195591] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.195883] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5f6cd28-5109-43d9-8971-dc1e944ada35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.210132] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-df0a9a69-bd44-4da4-ba3a-9ba241c010a4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.263601] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.263983] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.264238] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore2] f73b99bf-6a2b-4f21-b855-74965c95ed76 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.264567] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78fdc563-2feb-48e9-8a28-088c732782f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.272884] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 916.272884] env[61964]: value = "task-1041269" [ 916.272884] env[61964]: _type = "Task" [ 916.272884] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.282220] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041269, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.341052] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 34e00bc8-8124-4ee9-a0a4-c70bf611996c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 916.459716] env[61964]: INFO nova.compute.manager [-] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Took 1.03 seconds to deallocate network for instance. [ 916.594988] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041267, 'name': Rename_Task, 'duration_secs': 0.183898} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.597354] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.597616] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65d1a2f8-de73-457c-922a-cbcaab22e7cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.605515] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 916.605515] env[61964]: value = "task-1041270" [ 916.605515] env[61964]: _type = "Task" [ 916.605515] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.617414] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041270, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.644017] env[61964]: DEBUG nova.network.neutron [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.644296] env[61964]: DEBUG nova.network.neutron [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.783153] env[61964]: DEBUG oslo_vmware.api [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041269, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160235} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.783443] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.783634] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.783812] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.802533] env[61964]: INFO nova.scheduler.client.report [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance f73b99bf-6a2b-4f21-b855-74965c95ed76 [ 916.846412] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 58f1ad74-ea72-45c8-b98c-c85674989342 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 916.846862] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 916.846862] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 916.923127] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.923460] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.967409] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.117217] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041270, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.147029] env[61964]: DEBUG oslo_concurrency.lockutils [req-cbad362c-44a3-4d1d-9005-a4d8ddbda433 req-7cd0b79b-3253-4d50-b374-eb57cbc15ba2 service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.208642] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bf0375-6db0-434c-b47f-b3ff41e7ef10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.216618] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.216956] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4954d2b8-b0ab-492b-8996-a3c305ddc239 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.219438] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d45f1cb-f1e9-4307-9712-70253df10a45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.253412] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c46c13-2ef5-4148-88c9-b0d5d763ee1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.256127] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 917.256127] env[61964]: value = "task-1041271" [ 917.256127] env[61964]: _type = "Task" [ 917.256127] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.262615] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebef498-59f0-4766-a9c0-74075172f130 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.269250] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041271, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.278897] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.307291] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.426111] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 917.603653] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.617696] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041270, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.766829] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041271, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.781953] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 917.946439] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.117068] env[61964]: DEBUG oslo_vmware.api [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041270, 'name': PowerOnVM_Task, 'duration_secs': 1.06553} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.117301] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.117507] env[61964]: INFO nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Took 9.69 seconds to spawn the instance on the hypervisor. [ 918.117693] env[61964]: DEBUG nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 918.118465] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e7811d-f02a-41d9-8dd5-e6d947975b03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.172161] env[61964]: DEBUG nova.compute.manager [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Received event network-changed-9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 918.172506] env[61964]: DEBUG nova.compute.manager [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Refreshing instance network info cache due to event network-changed-9e14aef8-cd53-45ad-9b3c-152397bc01da. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 918.172847] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] Acquiring lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.173025] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] Acquired lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.173256] env[61964]: DEBUG nova.network.neutron [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Refreshing network info cache for port 9e14aef8-cd53-45ad-9b3c-152397bc01da {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.267246] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041271, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.287123] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 918.287386] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.503s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.287676] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.980s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.289202] env[61964]: INFO nova.compute.claims [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.636853] env[61964]: INFO nova.compute.manager [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Took 35.98 seconds to build instance. [ 918.767534] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041271, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.884068] env[61964]: DEBUG nova.network.neutron [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updated VIF entry in instance network info cache for port 9e14aef8-cd53-45ad-9b3c-152397bc01da. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.884463] env[61964]: DEBUG nova.network.neutron [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updating instance_info_cache with network_info: [{"id": "9e14aef8-cd53-45ad-9b3c-152397bc01da", "address": "fa:16:3e:b8:3b:74", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": null, "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9e14aef8-cd", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.139156] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e69e0fa8-fee3-4fc6-a403-fb29de85ed8f tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.490s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.268758] env[61964]: DEBUG oslo_vmware.api [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041271, 'name': PowerOnVM_Task, 'duration_secs': 2.002391} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.269098] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.269284] env[61964]: DEBUG nova.compute.manager [None req-d275e786-2cd4-4bdf-b4ef-70420c6374ee tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 919.269985] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d213a527-0f58-4032-bc35-a5bccfcccc2f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.387017] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ce424e5-4779-4d39-a91a-0c85e866f05e req-deac3c60-89c4-43e4-8d9c-2520629e6092 service nova] Releasing lock "refresh_cache-f73b99bf-6a2b-4f21-b855-74965c95ed76" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.559394] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be942041-6d1a-459b-85ef-115c93184e74 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.568189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185dd504-44f6-41e9-99a7-37f2c09aa071 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.599230] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914a40da-7120-4811-b49f-b3dc58406f58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.608117] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628c6cb9-582a-45e0-8a04-e8000229f7b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.621448] env[61964]: DEBUG nova.compute.provider_tree [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.124487] env[61964]: DEBUG nova.scheduler.client.report [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.200688] env[61964]: DEBUG nova.compute.manager [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Received event network-changed-d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 920.200922] env[61964]: DEBUG nova.compute.manager [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Refreshing instance network info cache due to event network-changed-d9618c13-facd-40f7-b201-776d03dffe57. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 920.201111] env[61964]: DEBUG oslo_concurrency.lockutils [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] Acquiring lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.201274] env[61964]: DEBUG oslo_concurrency.lockutils [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] Acquired lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.201500] env[61964]: DEBUG nova.network.neutron [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Refreshing network info cache for port d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.630282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.630875] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 920.633693] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.109s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.635213] env[61964]: INFO nova.compute.claims [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.826778] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.827046] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.827258] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.827448] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.827712] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.829779] env[61964]: INFO nova.compute.manager [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Terminating instance [ 920.940847] env[61964]: DEBUG nova.network.neutron [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updated VIF entry in instance network info cache for port d9618c13-facd-40f7-b201-776d03dffe57. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 920.941264] env[61964]: DEBUG nova.network.neutron [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating instance_info_cache with network_info: [{"id": "d9618c13-facd-40f7-b201-776d03dffe57", "address": "fa:16:3e:5c:78:03", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9618c13-fa", "ovs_interfaceid": "d9618c13-facd-40f7-b201-776d03dffe57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.140624] env[61964]: DEBUG nova.compute.utils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.143868] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 921.144084] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 921.189192] env[61964]: DEBUG nova.policy [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cb7b5913f4f44c6b227d47a61b8740a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd511575612441cc92145aa768353f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 921.334272] env[61964]: DEBUG nova.compute.manager [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 921.334648] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 921.335488] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdebc6cc-9f06-450f-9071-1f660b04d96a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.344482] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 921.344742] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a01b6cb-81e3-428b-9758-7e0e372c5f6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.353101] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 921.353101] env[61964]: value = "task-1041272" [ 921.353101] env[61964]: _type = "Task" [ 921.353101] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.362526] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.445972] env[61964]: DEBUG oslo_concurrency.lockutils [req-2475f7a6-b94c-4614-85b8-5a360fa153d6 req-72389fb3-6f19-47f9-ad0a-cfcb06e79fd8 service nova] Releasing lock "refresh_cache-8a14232f-5cc6-4856-9291-eb0a1bd2ab69" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.448496] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Successfully created port: c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.645276] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 921.863811] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041272, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.940475] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe595448-a8e7-4052-8a06-bfe9243277d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.950137] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc83010f-08f7-4267-9459-b1ae160aa4f5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.983682] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0038129-c056-42f8-8a9b-5bec086688c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.992320] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b2c056-94dc-46b0-be16-d5195f3f8d3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.011236] env[61964]: DEBUG nova.compute.provider_tree [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.365564] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041272, 'name': PowerOffVM_Task, 'duration_secs': 0.536364} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.365910] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 922.366078] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.366401] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4c0f692-8f3f-4381-ac7d-f859074f80c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.452516] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.452756] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.452956] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleting the datastore file [datastore1] df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.453246] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e2fd426-a5c6-4eba-92d5-5377ed9f3877 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.460796] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 922.460796] env[61964]: value = "task-1041274" [ 922.460796] env[61964]: _type = "Task" [ 922.460796] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.469625] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.515093] env[61964]: DEBUG nova.scheduler.client.report [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 922.657658] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 922.684142] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 922.684418] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.684808] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 922.685079] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.685237] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 922.685394] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 922.685610] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 922.685824] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 922.686052] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 922.686233] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 922.686416] env[61964]: DEBUG nova.virt.hardware [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 922.687289] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d637367-4ed4-4520-a96f-8878767182f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.696151] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbf4f78-b0e9-4cde-9635-9a7014dd9090 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.971496] env[61964]: DEBUG oslo_vmware.api [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.370941} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.971774] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.972018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.972240] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.972443] env[61964]: INFO nova.compute.manager [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 1.64 seconds to destroy the instance on the hypervisor. [ 922.972701] env[61964]: DEBUG oslo.service.loopingcall [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.972911] env[61964]: DEBUG nova.compute.manager [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 922.973033] env[61964]: DEBUG nova.network.neutron [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.993538] env[61964]: DEBUG nova.compute.manager [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Received event network-vif-plugged-c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 922.993771] env[61964]: DEBUG oslo_concurrency.lockutils [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] Acquiring lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.993864] env[61964]: DEBUG oslo_concurrency.lockutils [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.994091] env[61964]: DEBUG oslo_concurrency.lockutils [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.994285] env[61964]: DEBUG nova.compute.manager [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] No waiting events found dispatching network-vif-plugged-c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 922.994548] env[61964]: WARNING nova.compute.manager [req-96553c52-8475-4943-94d9-41544d74bbc7 req-c1387ea0-0f02-41e6-9b64-cba5dcca92c4 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Received unexpected event network-vif-plugged-c820ead7-034a-450f-839a-9bca2af7bb43 for instance with vm_state building and task_state spawning. [ 923.021114] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.021622] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 923.024157] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.106s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.025534] env[61964]: INFO nova.compute.claims [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.099434] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Successfully updated port: c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.531824] env[61964]: DEBUG nova.compute.utils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.533364] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 923.533649] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.556273] env[61964]: DEBUG nova.compute.manager [req-6ffc949c-d1c6-443a-acf6-c1ffe9ec0135 req-28b5a79d-5606-4c2a-a423-58079337cced service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Received event network-vif-deleted-8b2c44a6-df10-43fa-9afa-85167edbb650 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 923.556273] env[61964]: INFO nova.compute.manager [req-6ffc949c-d1c6-443a-acf6-c1ffe9ec0135 req-28b5a79d-5606-4c2a-a423-58079337cced service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Neutron deleted interface 8b2c44a6-df10-43fa-9afa-85167edbb650; detaching it from the instance and deleting it from the info cache [ 923.556273] env[61964]: DEBUG nova.network.neutron [req-6ffc949c-d1c6-443a-acf6-c1ffe9ec0135 req-28b5a79d-5606-4c2a-a423-58079337cced service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.575650] env[61964]: DEBUG nova.policy [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cb7b5913f4f44c6b227d47a61b8740a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd511575612441cc92145aa768353f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.602017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.602199] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.602357] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.689439] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.689836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.830057] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Successfully created port: 6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.029089] env[61964]: DEBUG nova.network.neutron [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.037073] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 924.060119] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8273dc8e-f56b-475e-8243-3a9fff979b70 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.069061] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cfc02f-bc17-40d8-b874-4141c7cbbdf3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.118031] env[61964]: DEBUG nova.compute.manager [req-6ffc949c-d1c6-443a-acf6-c1ffe9ec0135 req-28b5a79d-5606-4c2a-a423-58079337cced service nova] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Detach interface failed, port_id=8b2c44a6-df10-43fa-9afa-85167edbb650, reason: Instance df0a9a69-bd44-4da4-ba3a-9ba241c010a4 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 924.158985] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 924.193432] env[61964]: INFO nova.compute.manager [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Detaching volume 090cdd91-5f65-47f9-a5fb-3e8504507812 [ 924.242490] env[61964]: INFO nova.virt.block_device [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Attempting to driver detach volume 090cdd91-5f65-47f9-a5fb-3e8504507812 from mountpoint /dev/sdb [ 924.242736] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 924.242929] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230511', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'name': 'volume-090cdd91-5f65-47f9-a5fb-3e8504507812', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '138f44e4-e12e-4f89-a9b2-8a512b53cdf5', 'attached_at': '', 'detached_at': '', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'serial': '090cdd91-5f65-47f9-a5fb-3e8504507812'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 924.244090] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b965cb-efc8-4aa6-9890-2086e922a31c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.270557] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdd9c08-a685-4ab8-8bec-fd5a2a792955 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.280837] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d243b39-47c3-4097-beae-f028d9c69961 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.309118] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4180cab-2eed-482b-a93d-bf58304ce740 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.329277] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] The volume has not been displaced from its original location: [datastore2] volume-090cdd91-5f65-47f9-a5fb-3e8504507812/volume-090cdd91-5f65-47f9-a5fb-3e8504507812.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 924.334092] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 924.337043] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fe1a3e2-1b38-412a-bce6-d5558cea3649 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.359680] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 924.359680] env[61964]: value = "task-1041275" [ 924.359680] env[61964]: _type = "Task" [ 924.359680] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.372677] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.407215] env[61964]: DEBUG nova.network.neutron [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updating instance_info_cache with network_info: [{"id": "c820ead7-034a-450f-839a-9bca2af7bb43", "address": "fa:16:3e:dc:4b:85", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc820ead7-03", "ovs_interfaceid": "c820ead7-034a-450f-839a-9bca2af7bb43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.473183] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3013ca-41e8-477d-ab30-f4efa8d8cd19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.481816] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa72e6d-3139-438d-b7e0-0e70f778c81f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.513604] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3b74c5-6874-45f7-84f5-08dc957e1197 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.523212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe9b431-5403-474e-943c-9fe12e9de5cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.537861] env[61964]: INFO nova.compute.manager [-] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Took 1.56 seconds to deallocate network for instance. [ 924.538389] env[61964]: DEBUG nova.compute.provider_tree [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.870565] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041275, 'name': ReconfigVM_Task, 'duration_secs': 0.265946} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.870841] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 924.875579] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e765c6a-d70d-49ad-8b00-56d1fa4ed144 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.891459] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 924.891459] env[61964]: value = "task-1041276" [ 924.891459] env[61964]: _type = "Task" [ 924.891459] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.900276] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.910979] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.911314] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Instance network_info: |[{"id": "c820ead7-034a-450f-839a-9bca2af7bb43", "address": "fa:16:3e:dc:4b:85", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc820ead7-03", "ovs_interfaceid": "c820ead7-034a-450f-839a-9bca2af7bb43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 924.911764] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:4b:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c820ead7-034a-450f-839a-9bca2af7bb43', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.919591] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Creating folder: Project (fd511575612441cc92145aa768353f92). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 924.919891] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29dc3d95-c75a-4ede-9708-486b2edab741 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.932104] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Created folder: Project (fd511575612441cc92145aa768353f92) in parent group-v230360. [ 924.932348] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Creating folder: Instances. Parent ref: group-v230528. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 924.932637] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-079a27ab-b90d-45df-9a8c-310555104e13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.944028] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Created folder: Instances in parent group-v230528. [ 924.944028] env[61964]: DEBUG oslo.service.loopingcall [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.944028] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.944028] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88c90392-af66-42ca-a627-9f60addacf4e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.965048] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.965048] env[61964]: value = "task-1041279" [ 924.965048] env[61964]: _type = "Task" [ 924.965048] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.972963] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041279, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.021202] env[61964]: DEBUG nova.compute.manager [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Received event network-changed-c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 925.021348] env[61964]: DEBUG nova.compute.manager [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Refreshing instance network info cache due to event network-changed-c820ead7-034a-450f-839a-9bca2af7bb43. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 925.021599] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] Acquiring lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.021752] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] Acquired lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.021918] env[61964]: DEBUG nova.network.neutron [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Refreshing network info cache for port c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.042035] env[61964]: DEBUG nova.scheduler.client.report [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 925.050026] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 925.050729] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.080304] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 925.080560] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.080797] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 925.081030] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.081693] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 925.081693] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 925.081693] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 925.081693] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 925.081893] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 925.082038] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 925.082300] env[61964]: DEBUG nova.virt.hardware [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 925.083186] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ae96c0-b293-4100-859a-ea7b689ea9d4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.093042] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6e981b-342a-42f3-ad47-6c0da3b715e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.355115] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Successfully updated port: 6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.403036] env[61964]: DEBUG oslo_vmware.api [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041276, 'name': ReconfigVM_Task, 'duration_secs': 0.153882} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.403036] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230511', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'name': 'volume-090cdd91-5f65-47f9-a5fb-3e8504507812', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '138f44e4-e12e-4f89-a9b2-8a512b53cdf5', 'attached_at': '', 'detached_at': '', 'volume_id': '090cdd91-5f65-47f9-a5fb-3e8504507812', 'serial': '090cdd91-5f65-47f9-a5fb-3e8504507812'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 925.475999] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041279, 'name': CreateVM_Task, 'duration_secs': 0.326324} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.475999] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.476204] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.476292] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.476615] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.476865] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31c343b7-7e96-45c9-a29c-007df11a0f52 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.482213] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 925.482213] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52145fe7-bbcd-954d-4251-3b817c9c2f7a" [ 925.482213] env[61964]: _type = "Task" [ 925.482213] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.490787] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52145fe7-bbcd-954d-4251-3b817c9c2f7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.548611] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.548611] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 925.551692] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.258s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.551692] env[61964]: DEBUG nova.objects.instance [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'pci_requests' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.672378] env[61964]: DEBUG nova.compute.manager [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Received event network-vif-plugged-6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 925.672378] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Acquiring lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.672860] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.673195] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.673482] env[61964]: DEBUG nova.compute.manager [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] No waiting events found dispatching network-vif-plugged-6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 925.673880] env[61964]: WARNING nova.compute.manager [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Received unexpected event network-vif-plugged-6b326930-a132-4e0d-831c-c44b7ac08c9f for instance with vm_state building and task_state spawning. [ 925.674178] env[61964]: DEBUG nova.compute.manager [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Received event network-changed-6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 925.674466] env[61964]: DEBUG nova.compute.manager [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Refreshing instance network info cache due to event network-changed-6b326930-a132-4e0d-831c-c44b7ac08c9f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 925.674773] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Acquiring lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.675047] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Acquired lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.675705] env[61964]: DEBUG nova.network.neutron [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Refreshing network info cache for port 6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.779716] env[61964]: DEBUG nova.network.neutron [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updated VIF entry in instance network info cache for port c820ead7-034a-450f-839a-9bca2af7bb43. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.780126] env[61964]: DEBUG nova.network.neutron [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updating instance_info_cache with network_info: [{"id": "c820ead7-034a-450f-839a-9bca2af7bb43", "address": "fa:16:3e:dc:4b:85", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc820ead7-03", "ovs_interfaceid": "c820ead7-034a-450f-839a-9bca2af7bb43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.857687] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.945939] env[61964]: DEBUG nova.objects.instance [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.994034] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52145fe7-bbcd-954d-4251-3b817c9c2f7a, 'name': SearchDatastore_Task, 'duration_secs': 0.011035} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.994154] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.994350] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.994593] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.994779] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.994986] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.995278] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca88abb9-1bf8-43a8-8b09-66457011e899 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.003843] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.004036] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 926.004741] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59fd519d-2442-4fa9-afdb-b204f9118fba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.009684] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 926.009684] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5256d2a5-fb9a-0e4e-6c3f-861292a1a8bc" [ 926.009684] env[61964]: _type = "Task" [ 926.009684] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.017419] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5256d2a5-fb9a-0e4e-6c3f-861292a1a8bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.055308] env[61964]: DEBUG nova.compute.utils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.058390] env[61964]: DEBUG nova.objects.instance [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'numa_topology' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.058855] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 926.059090] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.099031] env[61964]: DEBUG nova.policy [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cb7b5913f4f44c6b227d47a61b8740a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd511575612441cc92145aa768353f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.216806] env[61964]: DEBUG nova.network.neutron [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.282890] env[61964]: DEBUG oslo_concurrency.lockutils [req-df4bb4af-a8ca-4ec2-b163-0d2b69617dde req-b9679a08-4eaa-4dc1-9bef-4be150c22abd service nova] Releasing lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.294402] env[61964]: DEBUG nova.network.neutron [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.411749] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Successfully created port: 9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.520911] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5256d2a5-fb9a-0e4e-6c3f-861292a1a8bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008065} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.521725] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04378a66-5730-4c9e-b8e0-3997d1e2ee43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.528550] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 926.528550] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbf16-ff93-5968-d90e-5ef468a44532" [ 926.528550] env[61964]: _type = "Task" [ 926.528550] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.536327] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbf16-ff93-5968-d90e-5ef468a44532, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.560073] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 926.563489] env[61964]: INFO nova.compute.claims [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.798023] env[61964]: DEBUG oslo_concurrency.lockutils [req-717f26e4-03e1-42c6-9e4d-f35b7a9265dc req-cf4162b0-26f7-4427-8b26-06b657cf20cc service nova] Releasing lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.798023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.798182] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 926.897108] env[61964]: DEBUG oslo_concurrency.lockutils [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.954047] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ff4e9a16-254f-4a4f-b22a-38f65d23d298 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.264s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.955019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.058s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.955235] env[61964]: DEBUG nova.compute.manager [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 926.956391] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875446f3-df22-42e3-969a-06972e677679 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.966457] env[61964]: DEBUG nova.compute.manager [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 926.966650] env[61964]: DEBUG nova.objects.instance [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.040964] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520bbf16-ff93-5968-d90e-5ef468a44532, 'name': SearchDatastore_Task, 'duration_secs': 0.009634} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.041274] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.041537] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] f5acabe4-674b-4c4c-85b4-b9270926017b/f5acabe4-674b-4c4c-85b4-b9270926017b.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 927.041827] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60b97a32-1a29-4025-98ad-91d4ec9ec103 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.050970] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 927.050970] env[61964]: value = "task-1041280" [ 927.050970] env[61964]: _type = "Task" [ 927.050970] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.058920] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.338339] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.516730] env[61964]: DEBUG nova.network.neutron [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Updating instance_info_cache with network_info: [{"id": "6b326930-a132-4e0d-831c-c44b7ac08c9f", "address": "fa:16:3e:a2:6f:1f", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b326930-a1", "ovs_interfaceid": "6b326930-a132-4e0d-831c-c44b7ac08c9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.561450] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041280, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.571217] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 927.600431] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 927.600759] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.600996] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 927.601275] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.601529] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 927.601751] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 927.601988] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 927.602173] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 927.602352] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 927.602522] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 927.603010] env[61964]: DEBUG nova.virt.hardware [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 927.604028] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245f8d97-21af-4ec4-abbf-494a7f6c5351 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.614726] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3634803-55a5-4084-a628-5109cf6f5f78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.834520] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779984d0-c87e-4113-bf45-d088bc884bd8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.842598] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce6c1f5-da01-4f37-bd6e-cf89db47161d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.874851] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c56298f-c1af-4833-a493-25b3c7ce82ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.884219] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c41c574-12e3-49da-9830-c80f5257b659 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.899831] env[61964]: DEBUG nova.compute.provider_tree [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.968196] env[61964]: DEBUG nova.compute.manager [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Received event network-vif-plugged-9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 927.968495] env[61964]: DEBUG oslo_concurrency.lockutils [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] Acquiring lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.968653] env[61964]: DEBUG oslo_concurrency.lockutils [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] Lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.968792] env[61964]: DEBUG oslo_concurrency.lockutils [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] Lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.968964] env[61964]: DEBUG nova.compute.manager [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] No waiting events found dispatching network-vif-plugged-9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 927.969332] env[61964]: WARNING nova.compute.manager [req-3818f0a8-62d6-4f7f-b708-e0b76f0d4015 req-4d6f861e-7ebf-4541-8b11-50226bcbb430 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Received unexpected event network-vif-plugged-9420c5ba-2284-4fb8-a95e-74558e7d7a9b for instance with vm_state building and task_state spawning. [ 927.973511] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 927.974355] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d74e42a-ae48-4d7d-895f-c26e6a83880a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.982092] env[61964]: DEBUG oslo_vmware.api [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 927.982092] env[61964]: value = "task-1041281" [ 927.982092] env[61964]: _type = "Task" [ 927.982092] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.991482] env[61964]: DEBUG oslo_vmware.api [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041281, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.019668] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "refresh_cache-003a5b74-2b8d-4e14-a7ee-db8006f81dfa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.019954] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance network_info: |[{"id": "6b326930-a132-4e0d-831c-c44b7ac08c9f", "address": "fa:16:3e:a2:6f:1f", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b326930-a1", "ovs_interfaceid": "6b326930-a132-4e0d-831c-c44b7ac08c9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 928.020370] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:6f:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b326930-a132-4e0d-831c-c44b7ac08c9f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.028023] env[61964]: DEBUG oslo.service.loopingcall [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.028207] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 928.028436] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc8adda2-4ee3-478c-a432-b20f0bee3261 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.049735] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.049735] env[61964]: value = "task-1041282" [ 928.049735] env[61964]: _type = "Task" [ 928.049735] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.066035] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041282, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.066349] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560162} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.066583] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] f5acabe4-674b-4c4c-85b4-b9270926017b/f5acabe4-674b-4c4c-85b4-b9270926017b.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.066796] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.067055] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-67eed1ad-7b7b-4473-8762-60636d9248f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.073406] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Successfully updated port: 9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.075770] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 928.075770] env[61964]: value = "task-1041283" [ 928.075770] env[61964]: _type = "Task" [ 928.075770] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.086954] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041283, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.403649] env[61964]: DEBUG nova.scheduler.client.report [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.492236] env[61964]: DEBUG oslo_vmware.api [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041281, 'name': PowerOffVM_Task, 'duration_secs': 0.248033} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.492778] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 928.492987] env[61964]: DEBUG nova.compute.manager [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.493783] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2face65f-6f71-4bc1-b498-352b0babb491 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.559696] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041282, 'name': CreateVM_Task, 'duration_secs': 0.392192} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.559865] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 928.560569] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.560740] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.561075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.561329] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-683622a8-c17d-4006-924b-0b4afe4e5013 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.565806] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 928.565806] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526d63af-db23-c6a3-74d5-0bef8b0c4500" [ 928.565806] env[61964]: _type = "Task" [ 928.565806] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.573526] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526d63af-db23-c6a3-74d5-0bef8b0c4500, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.577109] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.577240] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.577388] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.585993] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041283, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080658} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.586770] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.587549] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaaf5465-c69f-4d8f-8624-ee72215d9d04 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.609612] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] f5acabe4-674b-4c4c-85b4-b9270926017b/f5acabe4-674b-4c4c-85b4-b9270926017b.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.610143] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e52c45e-2a5c-4987-9b6f-f6bf41f71d68 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.630387] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 928.630387] env[61964]: value = "task-1041284" [ 928.630387] env[61964]: _type = "Task" [ 928.630387] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.638226] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041284, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.909073] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.358s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.911997] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 24.079s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.972287] env[61964]: INFO nova.network.neutron [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating port edda620f-1259-4fbb-afa6-48aef4eda40b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 929.005942] env[61964]: DEBUG oslo_concurrency.lockutils [None req-89c5c86d-762e-4891-8421-6b3b7a96ac7d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.077572] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526d63af-db23-c6a3-74d5-0bef8b0c4500, 'name': SearchDatastore_Task, 'duration_secs': 0.010689} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.077921] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.078178] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.078420] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.078564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.078760] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.079056] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efc8514a-5c88-4a80-b895-1c7e1ed70231 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.088229] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.088438] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 929.089160] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39df27bc-4879-451e-9983-884efa215f31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.095202] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 929.095202] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba0702-2b77-b0b3-85c1-a48ad12818f6" [ 929.095202] env[61964]: _type = "Task" [ 929.095202] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.103631] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba0702-2b77-b0b3-85c1-a48ad12818f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.117927] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.140532] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041284, 'name': ReconfigVM_Task, 'duration_secs': 0.315791} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.142922] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Reconfigured VM instance instance-00000054 to attach disk [datastore1] f5acabe4-674b-4c4c-85b4-b9270926017b/f5acabe4-674b-4c4c-85b4-b9270926017b.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.143602] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7f78824-32ef-4891-a708-1df3161c0e63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.152142] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 929.152142] env[61964]: value = "task-1041285" [ 929.152142] env[61964]: _type = "Task" [ 929.152142] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.161056] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041285, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.307882] env[61964]: DEBUG nova.network.neutron [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Updating instance_info_cache with network_info: [{"id": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "address": "fa:16:3e:b0:dc:5b", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9420c5ba-22", "ovs_interfaceid": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.410408] env[61964]: DEBUG nova.objects.instance [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.605786] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ba0702-2b77-b0b3-85c1-a48ad12818f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009053} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.607026] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-660a27a0-4a73-4c66-b18e-2e6e5d07966a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.614205] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 929.614205] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520c623e-3774-41a9-645a-6b5484e5ffee" [ 929.614205] env[61964]: _type = "Task" [ 929.614205] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.622138] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520c623e-3774-41a9-645a-6b5484e5ffee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.656988] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16b5794-577a-472e-9683-9a3a199994f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.666624] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041285, 'name': Rename_Task, 'duration_secs': 0.208251} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.666953] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 929.667896] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1bb53b-e45c-4b73-b6cd-988126a1e3d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.670725] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08be2b63-dde2-47d4-bb12-a78f9db5274e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.700897] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708a4fe0-7155-44d5-b58d-c02b012934c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.703414] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 929.703414] env[61964]: value = "task-1041286" [ 929.703414] env[61964]: _type = "Task" [ 929.703414] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.709801] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44fb833-575d-4e50-892e-9618fe0de21f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.716549] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041286, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.727055] env[61964]: DEBUG nova.compute.provider_tree [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.810378] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.810699] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Instance network_info: |[{"id": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "address": "fa:16:3e:b0:dc:5b", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9420c5ba-22", "ovs_interfaceid": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 929.811141] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:dc:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9420c5ba-2284-4fb8-a95e-74558e7d7a9b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.818568] env[61964]: DEBUG oslo.service.loopingcall [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.818767] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 929.818987] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44f9945f-38cf-4088-943f-7d0219d8e34e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.839114] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.839114] env[61964]: value = "task-1041287" [ 929.839114] env[61964]: _type = "Task" [ 929.839114] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.846812] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041287, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.916617] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.916997] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.917365] env[61964]: DEBUG nova.network.neutron [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.917718] env[61964]: DEBUG nova.objects.instance [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'info_cache' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.993444] env[61964]: DEBUG nova.compute.manager [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Received event network-changed-9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 929.993658] env[61964]: DEBUG nova.compute.manager [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Refreshing instance network info cache due to event network-changed-9420c5ba-2284-4fb8-a95e-74558e7d7a9b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 929.993905] env[61964]: DEBUG oslo_concurrency.lockutils [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] Acquiring lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.994109] env[61964]: DEBUG oslo_concurrency.lockutils [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] Acquired lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.994303] env[61964]: DEBUG nova.network.neutron [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Refreshing network info cache for port 9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.132104] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520c623e-3774-41a9-645a-6b5484e5ffee, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.132104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.132104] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 003a5b74-2b8d-4e14-a7ee-db8006f81dfa/003a5b74-2b8d-4e14-a7ee-db8006f81dfa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 930.132104] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1322caec-f719-4aac-8b3f-a7b684d25ffa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.141070] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 930.141070] env[61964]: value = "task-1041288" [ 930.141070] env[61964]: _type = "Task" [ 930.141070] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.151794] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041288, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.214242] env[61964]: DEBUG oslo_vmware.api [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041286, 'name': PowerOnVM_Task, 'duration_secs': 0.494866} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.214549] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 930.214784] env[61964]: INFO nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Took 7.56 seconds to spawn the instance on the hypervisor. [ 930.214996] env[61964]: DEBUG nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 930.215885] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458d1a0c-f228-48e1-96b7-552e4630bebd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.230909] env[61964]: DEBUG nova.scheduler.client.report [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.350664] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041287, 'name': CreateVM_Task, 'duration_secs': 0.329305} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.350856] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.351662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.351871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.352263] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.352575] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd894cc2-928e-421c-9872-47c479a4541a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.359644] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 930.359644] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c1b25-450a-e8cc-5cb2-09a3468ae410" [ 930.359644] env[61964]: _type = "Task" [ 930.359644] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.369667] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c1b25-450a-e8cc-5cb2-09a3468ae410, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.422197] env[61964]: DEBUG nova.objects.base [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Object Instance<138f44e4-e12e-4f89-a9b2-8a512b53cdf5> lazy-loaded attributes: flavor,info_cache {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 930.653437] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041288, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447748} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.653766] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 003a5b74-2b8d-4e14-a7ee-db8006f81dfa/003a5b74-2b8d-4e14-a7ee-db8006f81dfa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 930.654018] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.654300] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d870df02-4feb-4db3-a35b-9c2ee671135a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.661734] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 930.661734] env[61964]: value = "task-1041289" [ 930.661734] env[61964]: _type = "Task" [ 930.661734] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.672315] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041289, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.714874] env[61964]: DEBUG nova.network.neutron [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Updated VIF entry in instance network info cache for port 9420c5ba-2284-4fb8-a95e-74558e7d7a9b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 930.715253] env[61964]: DEBUG nova.network.neutron [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Updating instance_info_cache with network_info: [{"id": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "address": "fa:16:3e:b0:dc:5b", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9420c5ba-22", "ovs_interfaceid": "9420c5ba-2284-4fb8-a95e-74558e7d7a9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.734101] env[61964]: INFO nova.compute.manager [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Took 43.44 seconds to build instance. [ 930.870759] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c1b25-450a-e8cc-5cb2-09a3468ae410, 'name': SearchDatastore_Task, 'duration_secs': 0.050151} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.871048] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.871325] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.871570] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.871720] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.871930] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.872212] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9a5f7f1-29c2-49f7-be3e-bee7144d76ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.880862] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.881074] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 930.881845] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3e02596-deb1-4c71-91d8-9bdfbb718c53 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.888081] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 930.888081] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c14d72-8998-8ea7-7a00-d91a9ae1af56" [ 930.888081] env[61964]: _type = "Task" [ 930.888081] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.896242] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c14d72-8998-8ea7-7a00-d91a9ae1af56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.131060] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.131060] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.131060] env[61964]: DEBUG nova.network.neutron [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.168634] env[61964]: DEBUG nova.network.neutron [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [{"id": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "address": "fa:16:3e:d9:51:c2", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.233", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a986bcd-65", "ovs_interfaceid": "3a986bcd-6584-4440-bf01-7db0dff75b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.177065] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041289, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062213} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.177065] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.177275] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc3e842-b0e0-446c-b9c1-3b8fa6f1e7cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.200737] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 003a5b74-2b8d-4e14-a7ee-db8006f81dfa/003a5b74-2b8d-4e14-a7ee-db8006f81dfa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.201348] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-907ff494-cc21-40ad-b404-a8e4c284ed03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.217422] env[61964]: DEBUG oslo_concurrency.lockutils [req-addff8cc-ed9b-438d-a2d9-69ea7f19b169 req-0e81eeed-4ea3-464c-b16a-e390fe948694 service nova] Releasing lock "refresh_cache-d8f24177-b695-4d40-958d-9272b4f683cc" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.221788] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 931.221788] env[61964]: value = "task-1041290" [ 931.221788] env[61964]: _type = "Task" [ 931.221788] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.231202] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.240326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.328s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.243182] env[61964]: DEBUG oslo_concurrency.lockutils [None req-91a18658-0fba-4079-8c2c-f5fc39b4049e tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.977s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.243665] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.086s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.245345] env[61964]: INFO nova.compute.claims [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.400597] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c14d72-8998-8ea7-7a00-d91a9ae1af56, 'name': SearchDatastore_Task, 'duration_secs': 0.009997} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.400949] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142cfdc8-b179-43bb-9fa0-41b509a5f28a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.412018] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 931.412018] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527ee1ae-0a7f-293a-859a-c67dc9de3c2c" [ 931.412018] env[61964]: _type = "Task" [ 931.412018] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.419973] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527ee1ae-0a7f-293a-859a-c67dc9de3c2c, 'name': SearchDatastore_Task} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.419973] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.420305] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8f24177-b695-4d40-958d-9272b4f683cc/d8f24177-b695-4d40-958d-9272b4f683cc.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 931.420571] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c16f4427-a1ad-4619-89df-744ea7ac141c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.428622] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 931.428622] env[61964]: value = "task-1041291" [ 931.428622] env[61964]: _type = "Task" [ 931.428622] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.437964] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.673276] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "refresh_cache-138f44e4-e12e-4f89-a9b2-8a512b53cdf5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.735096] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041290, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.817520] env[61964]: INFO nova.scheduler.client.report [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocation for migration 8b2cfb5a-d6e6-49e9-8d5f-74cdad8f363a [ 931.878904] env[61964]: DEBUG nova.network.neutron [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.940440] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041291, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.022685] env[61964]: DEBUG nova.compute.manager [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 932.022897] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.023731] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.024220] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.024220] env[61964]: DEBUG nova.compute.manager [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] No waiting events found dispatching network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 932.024334] env[61964]: WARNING nova.compute.manager [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received unexpected event network-vif-plugged-edda620f-1259-4fbb-afa6-48aef4eda40b for instance with vm_state shelved_offloaded and task_state spawning. [ 932.025038] env[61964]: DEBUG nova.compute.manager [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 932.025038] env[61964]: DEBUG nova.compute.manager [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing instance network info cache due to event network-changed-edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 932.025038] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.239536] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041290, 'name': ReconfigVM_Task, 'duration_secs': 0.824383} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.239829] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 003a5b74-2b8d-4e14-a7ee-db8006f81dfa/003a5b74-2b8d-4e14-a7ee-db8006f81dfa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.240475] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5d3f08f-4139-4734-a537-cdecf491a560 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.253023] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 932.253023] env[61964]: value = "task-1041292" [ 932.253023] env[61964]: _type = "Task" [ 932.253023] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.266488] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041292, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.323564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2b898fbe-17ff-4202-b7c2-ac94ac25aba7 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.974s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.381793] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.384524] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.384889] env[61964]: DEBUG nova.network.neutron [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Refreshing network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.414773] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='90bc92cd5329b142bd87e6e70d218481',container_format='bare',created_at=2024-09-30T16:25:38Z,direct_url=,disk_format='vmdk',id=f732a7b4-4004-478c-ad84-b0f50659c288,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1713350290-shelved',owner='b772e2152b674da8bdc2cb4f726f6772',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-09-30T16:25:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 932.415055] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.415221] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 932.415407] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.415558] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 932.415728] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 932.415969] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 932.416189] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 932.416334] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 932.416884] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 932.416884] env[61964]: DEBUG nova.virt.hardware [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 932.417634] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ae2148-f5bb-4e15-bd3f-aab61eb23f99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.429854] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9435ea6b-4a02-4314-abcb-563022ba0f9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.459115] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:0b:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '02092ea4-bae0-4e42-b0ab-abc365b4395a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'edda620f-1259-4fbb-afa6-48aef4eda40b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.465038] env[61964]: DEBUG oslo.service.loopingcall [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.465387] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533382} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.468324] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.468591] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] d8f24177-b695-4d40-958d-9272b4f683cc/d8f24177-b695-4d40-958d-9272b4f683cc.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 932.468796] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.469269] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c45acc43-64b4-4a84-a524-70d2d141d4a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.483853] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94a4003e-0cc3-46c7-b79c-0904bac90094 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.494950] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 932.494950] env[61964]: value = "task-1041293" [ 932.494950] env[61964]: _type = "Task" [ 932.494950] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.496218] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.496218] env[61964]: value = "task-1041294" [ 932.496218] env[61964]: _type = "Task" [ 932.496218] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.511920] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041294, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.518745] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.587351] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ed3530-e8dd-454b-9639-36a0a846506a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.602641] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04dcde81-59b9-49d4-bbb1-c474eb833ab5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.353269] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.355959] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.356159] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.356459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "84821045-27b2-4a99-87f9-988b8615b83f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.356580] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.356782] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.358085] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8566e7b-2631-465f-8269-3c5199b2a419 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.368036] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5ffe2d-a413-43d4-8472-1c3b4ac1c844 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.372969] env[61964]: INFO nova.compute.manager [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Terminating instance [ 933.383758] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084204} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.384121] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041292, 'name': Rename_Task, 'duration_secs': 0.359763} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.388592] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.388912] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.389116] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041294, 'name': CreateVM_Task, 'duration_secs': 0.396643} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.389377] env[61964]: DEBUG oslo_vmware.api [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 933.389377] env[61964]: value = "task-1041295" [ 933.389377] env[61964]: _type = "Task" [ 933.389377] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.390102] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3921522-93ca-40e5-b3ae-57cab9e69fac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.392676] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bdb0252-bbc3-45fa-ac16-048b81a25793 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.393871] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.396833] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.397044] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.397384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.400899] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae30270a-4adc-4e8c-bad5-ee27eb1670e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.403346] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843b8e4f-c4ec-4a4f-af06-82db4d8317dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.417127] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 933.417127] env[61964]: value = "task-1041296" [ 933.417127] env[61964]: _type = "Task" [ 933.417127] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.426095] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] d8f24177-b695-4d40-958d-9272b4f683cc/d8f24177-b695-4d40-958d-9272b4f683cc.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.430912] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69d464dc-64ef-4dc7-b3d5-4bf10d910ed3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.446595] env[61964]: DEBUG oslo_vmware.api [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041295, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.459644] env[61964]: DEBUG nova.compute.provider_tree [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.460958] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 933.460958] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f01009-fe29-3c86-4bcc-0168e310010b" [ 933.460958] env[61964]: _type = "Task" [ 933.460958] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.470914] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041296, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.472952] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 933.472952] env[61964]: value = "task-1041297" [ 933.472952] env[61964]: _type = "Task" [ 933.472952] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.480467] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.480765] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Processing image f732a7b4-4004-478c-ad84-b0f50659c288 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.481018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.481182] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.481465] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.482418] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-caec0397-a582-4018-87e4-bffb9c1bccd7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.489221] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041297, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.497370] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.497707] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 933.498447] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df299c4c-73f5-4efd-aa38-9f6b05088653 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.504770] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 933.504770] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52548010-fda1-435c-0a1d-6793cd60ffc6" [ 933.504770] env[61964]: _type = "Task" [ 933.504770] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.525396] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 933.525702] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Fetch image to [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7/OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 933.525934] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Downloading stream optimized image f732a7b4-4004-478c-ad84-b0f50659c288 to [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7/OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7.vmdk on the data store datastore2 as vApp {{(pid=61964) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 933.526077] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Downloading image file data f732a7b4-4004-478c-ad84-b0f50659c288 to the ESX as VM named 'OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7' {{(pid=61964) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 933.535737] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.535845] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.612084] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 933.612084] env[61964]: value = "resgroup-9" [ 933.612084] env[61964]: _type = "ResourcePool" [ 933.612084] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 933.612601] env[61964]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-22a758f9-f15e-4bfa-ac17-b2e84bceadbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.634076] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lease: (returnval){ [ 933.634076] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 933.634076] env[61964]: _type = "HttpNfcLease" [ 933.634076] env[61964]: } obtained for vApp import into resource pool (val){ [ 933.634076] env[61964]: value = "resgroup-9" [ 933.634076] env[61964]: _type = "ResourcePool" [ 933.634076] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 933.634401] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the lease: (returnval){ [ 933.634401] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 933.634401] env[61964]: _type = "HttpNfcLease" [ 933.634401] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 933.640625] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 933.640625] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 933.640625] env[61964]: _type = "HttpNfcLease" [ 933.640625] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 933.678441] env[61964]: DEBUG nova.network.neutron [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updated VIF entry in instance network info cache for port edda620f-1259-4fbb-afa6-48aef4eda40b. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.678820] env[61964]: DEBUG nova.network.neutron [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.885283] env[61964]: DEBUG nova.compute.manager [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 933.885575] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.886628] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7629ae4-6cfc-42d3-a719-6f95a9b87d3b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.896000] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.899736] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d79a45d-0284-4e32-9257-c4eb7810297a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.908243] env[61964]: DEBUG oslo_vmware.api [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041295, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.909851] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 933.909851] env[61964]: value = "task-1041299" [ 933.909851] env[61964]: _type = "Task" [ 933.909851] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.919676] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.941848] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041296, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.966051] env[61964]: DEBUG nova.scheduler.client.report [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.985419] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041297, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.038714] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 934.147062] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 934.147062] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 934.147062] env[61964]: _type = "HttpNfcLease" [ 934.147062] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 934.181510] env[61964]: DEBUG oslo_concurrency.lockutils [req-5876027a-52cd-4c8e-b615-0c25db15e053 req-67eeb888-e58b-4a81-9e39-640442c621a5 service nova] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.406368] env[61964]: DEBUG oslo_vmware.api [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041295, 'name': PowerOnVM_Task, 'duration_secs': 0.594403} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.406721] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.406840] env[61964]: DEBUG nova.compute.manager [None req-4982eb09-9a76-453c-b31f-00923e3ace3a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 934.407583] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b4b57f-7dce-49e6-bb48-32dd2f8234c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.421722] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041299, 'name': PowerOffVM_Task, 'duration_secs': 0.335033} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.421969] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.422169] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.422398] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92b6a546-f2fc-4323-b945-ecb5131b2c62 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.439830] env[61964]: DEBUG oslo_vmware.api [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041296, 'name': PowerOnVM_Task, 'duration_secs': 0.726037} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.440086] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 934.440319] env[61964]: INFO nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Took 9.39 seconds to spawn the instance on the hypervisor. [ 934.440502] env[61964]: DEBUG nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 934.441224] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf6de4d-0e55-4bf9-b433-c9a5d11c47f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.473017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.229s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.473566] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 934.476597] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.623s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.476712] env[61964]: DEBUG nova.objects.instance [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61964) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 934.489840] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041297, 'name': ReconfigVM_Task, 'duration_secs': 0.551248} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.490162] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Reconfigured VM instance instance-00000056 to attach disk [datastore1] d8f24177-b695-4d40-958d-9272b4f683cc/d8f24177-b695-4d40-958d-9272b4f683cc.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.490832] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f08a440c-bf3f-4e98-8c53-8fbf6b24d463 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.499033] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 934.499033] env[61964]: value = "task-1041301" [ 934.499033] env[61964]: _type = "Task" [ 934.499033] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.504026] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.504282] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.504559] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] 84821045-27b2-4a99-87f9-988b8615b83f {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.505367] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a34f3fa8-538c-4737-be69-58a6215446c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.510587] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041301, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.516104] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 934.516104] env[61964]: value = "task-1041302" [ 934.516104] env[61964]: _type = "Task" [ 934.516104] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.525884] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.562987] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.644277] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 934.644277] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 934.644277] env[61964]: _type = "HttpNfcLease" [ 934.644277] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 934.644601] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 934.644601] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f47477-978e-6d9d-93e0-f384525c3bbb" [ 934.644601] env[61964]: _type = "HttpNfcLease" [ 934.644601] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 934.645374] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50320d18-5b72-4ab8-bfc4-f8102bd502e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.653451] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 934.653643] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 934.719478] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9a8831d2-0213-4bed-9cd6-dac542e3323a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.959986] env[61964]: INFO nova.compute.manager [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Took 47.45 seconds to build instance. [ 934.981729] env[61964]: DEBUG nova.compute.utils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.986261] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 934.986261] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.015195] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041301, 'name': Rename_Task, 'duration_secs': 0.316017} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.015503] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 935.015762] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5af30e9a-701b-4b96-9fd6-912213f5b85b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.028667] env[61964]: DEBUG oslo_vmware.api [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34378} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.031865] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.032743] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.032743] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.032743] env[61964]: INFO nova.compute.manager [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 935.032743] env[61964]: DEBUG oslo.service.loopingcall [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.032985] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 935.032985] env[61964]: value = "task-1041303" [ 935.032985] env[61964]: _type = "Task" [ 935.032985] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.033181] env[61964]: DEBUG nova.compute.manager [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 935.033299] env[61964]: DEBUG nova.network.neutron [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.045688] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041303, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.048567] env[61964]: DEBUG nova.policy [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ba7586bbae64e5f9cf2724cf0893b13', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1658dcfca064b058aaac985b9c7a5de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.459318] env[61964]: DEBUG nova.compute.manager [req-2b855a4c-0c10-4e13-bcd7-d7a8240ed9bf req-06a51e84-7d25-4e18-8569-a4c519c42381 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Received event network-vif-deleted-2a884ac5-7364-4168-bf0a-d947f84b8b92 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 935.459600] env[61964]: INFO nova.compute.manager [req-2b855a4c-0c10-4e13-bcd7-d7a8240ed9bf req-06a51e84-7d25-4e18-8569-a4c519c42381 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Neutron deleted interface 2a884ac5-7364-4168-bf0a-d947f84b8b92; detaching it from the instance and deleting it from the info cache [ 935.459796] env[61964]: DEBUG nova.network.neutron [req-2b855a4c-0c10-4e13-bcd7-d7a8240ed9bf req-06a51e84-7d25-4e18-8569-a4c519c42381 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.465896] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1b28f8be-09fa-4d94-b568-7f114dec326a tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.963s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.493067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cd9a0aa3-65e5-445e-a0c8-5782b75a56d6 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.496301] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.408s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.497798] env[61964]: INFO nova.compute.claims [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.500993] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 935.553716] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041303, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.592192] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Successfully created port: d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.889124] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 935.889434] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 935.890472] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5818683-05bc-425a-a900-a756c99afe03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.898190] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 935.898426] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 935.898701] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-958b61b4-9c10-4d6a-b917-54f5da92d3ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.936043] env[61964]: DEBUG nova.network.neutron [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.967435] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d860d6b8-e929-415d-9a48-1335402b2bb5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.978494] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1981ab1a-0fd0-42fa-a6cc-afb7edcc41bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.016645] env[61964]: DEBUG nova.compute.manager [req-2b855a4c-0c10-4e13-bcd7-d7a8240ed9bf req-06a51e84-7d25-4e18-8569-a4c519c42381 service nova] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Detach interface failed, port_id=2a884ac5-7364-4168-bf0a-d947f84b8b92, reason: Instance 84821045-27b2-4a99-87f9-988b8615b83f could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 936.049367] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041303, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.100049] env[61964]: DEBUG oslo_vmware.rw_handles [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e331dc-9522-5eb1-3748-df8d60d89b2c/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 936.100049] env[61964]: INFO nova.virt.vmwareapi.images [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Downloaded image file data f732a7b4-4004-478c-ad84-b0f50659c288 [ 936.100434] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e755170e-fea1-408b-af44-aeb73cb553f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.119174] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef1b471f-8ff2-45dc-8998-b254a89c6bac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.156040] env[61964]: INFO nova.virt.vmwareapi.images [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] The imported VM was unregistered [ 936.160019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 936.160491] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Creating directory with path [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.160846] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19b900d4-7bfb-45cf-bc20-b2f931b43027 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.174271] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Created directory with path [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.174612] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7/OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7.vmdk to [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk. {{(pid=61964) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 936.174889] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-bc1b2bff-0302-4649-aae9-e1216637086d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.183671] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 936.183671] env[61964]: value = "task-1041305" [ 936.183671] env[61964]: _type = "Task" [ 936.183671] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.194612] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.439169] env[61964]: INFO nova.compute.manager [-] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Took 1.41 seconds to deallocate network for instance. [ 936.521963] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 936.551746] env[61964]: DEBUG oslo_vmware.api [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041303, 'name': PowerOnVM_Task, 'duration_secs': 1.327185} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.555277] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 936.555705] env[61964]: INFO nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Took 8.98 seconds to spawn the instance on the hypervisor. [ 936.555929] env[61964]: DEBUG nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 936.558846] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 936.559286] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.559475] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.559677] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.559829] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.559979] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 936.560425] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 936.560623] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 936.560812] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 936.560983] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 936.561214] env[61964]: DEBUG nova.virt.hardware [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 936.562259] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355390f1-ba24-407f-b699-b821c2cbf2da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.566043] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6b703d-bfe6-44ca-aaed-32149ee84ebc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.582776] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb33e7c4-f4d6-4997-988d-6d07ac20f234 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.696858] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.915975] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d56018b-d52e-4df4-81b3-71297ee42703 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.924978] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac9c778-d599-4787-ab54-3b9ae2b1f7c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.963881] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.965146] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbaa13d-70ed-4ffe-b322-a0bb2595f19c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.974876] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a3c365-3d82-4640-a626-c54363f93cd8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.992064] env[61964]: DEBUG nova.compute.provider_tree [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.091703] env[61964]: INFO nova.compute.manager [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Took 49.19 seconds to build instance. [ 937.095264] env[61964]: DEBUG nova.compute.manager [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Received event network-vif-plugged-d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 937.095682] env[61964]: DEBUG oslo_concurrency.lockutils [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] Acquiring lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.097620] env[61964]: DEBUG oslo_concurrency.lockutils [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.097620] env[61964]: DEBUG oslo_concurrency.lockutils [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.097620] env[61964]: DEBUG nova.compute.manager [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] No waiting events found dispatching network-vif-plugged-d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 937.097620] env[61964]: WARNING nova.compute.manager [req-2a352156-537a-4042-9a25-bc448be33e4e req-b4180879-53f6-43c6-b84e-08857a557ddc service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Received unexpected event network-vif-plugged-d6798518-eaec-483c-9a59-bcca91699812 for instance with vm_state building and task_state spawning. [ 937.197927] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.276557] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Successfully updated port: d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.495034] env[61964]: DEBUG nova.scheduler.client.report [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 937.594150] env[61964]: DEBUG oslo_concurrency.lockutils [None req-42d51ba0-45f4-4c47-a44b-9b9269daf5c6 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.697s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.704954] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.778476] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.778748] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquired lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.778920] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.000612] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.001649] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 938.004953] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.064s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.005224] env[61964]: DEBUG nova.objects.instance [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'resources' on Instance uuid 8bf586a2-0b56-4f5a-9f33-d537fba00d32 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.199678] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.321772] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.508142] env[61964]: DEBUG nova.compute.utils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.514320] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 938.514320] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 938.549549] env[61964]: DEBUG nova.network.neutron [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Updating instance_info_cache with network_info: [{"id": "d6798518-eaec-483c-9a59-bcca91699812", "address": "fa:16:3e:c4:4e:8d", "network": {"id": "68d4d727-79b9-4d39-a8e0-fa481cfa70c5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1233288901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1658dcfca064b058aaac985b9c7a5de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6798518-ea", "ovs_interfaceid": "d6798518-eaec-483c-9a59-bcca91699812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.580205] env[61964]: DEBUG nova.policy [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05bfff4e5dae488593812635d1c216b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3f1804a0ded74317a1e2016db18c55ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.704789] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041305, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.481297} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.705257] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7/OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7.vmdk to [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk. [ 938.705484] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Cleaning up location [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 938.705675] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_33104610-97dc-44f2-a02c-1d178910ebe7 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.708124] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51fc9b4e-2216-4b12-95ed-e4487e7f92c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.717135] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 938.717135] env[61964]: value = "task-1041306" [ 938.717135] env[61964]: _type = "Task" [ 938.717135] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.730227] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.836381] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f37bcba-350c-413b-9b51-52c4eb3903a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.846031] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a11165-2174-4db1-a4c4-12cb3ff783da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.881806] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd266770-355b-4093-81e9-0267214f1215 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.888014] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3921a262-dda6-4e78-9906-9472340fe7d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.902943] env[61964]: DEBUG nova.compute.provider_tree [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.908136] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Successfully created port: f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.013807] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 939.053581] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Releasing lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.055165] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Instance network_info: |[{"id": "d6798518-eaec-483c-9a59-bcca91699812", "address": "fa:16:3e:c4:4e:8d", "network": {"id": "68d4d727-79b9-4d39-a8e0-fa481cfa70c5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1233288901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1658dcfca064b058aaac985b9c7a5de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6798518-ea", "ovs_interfaceid": "d6798518-eaec-483c-9a59-bcca91699812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 939.056411] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:4e:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '96d8be6c-b557-4b40-b0f5-838c62a3c904', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6798518-eaec-483c-9a59-bcca91699812', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.068021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Creating folder: Project (c1658dcfca064b058aaac985b9c7a5de). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 939.068021] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0e7163e-984f-4407-a389-0688edab1990 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.078945] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Created folder: Project (c1658dcfca064b058aaac985b9c7a5de) in parent group-v230360. [ 939.079172] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Creating folder: Instances. Parent ref: group-v230535. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 939.079471] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-501e0d1e-8535-49bf-8eb4-0b5e22df9bdc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.091021] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Created folder: Instances in parent group-v230535. [ 939.091810] env[61964]: DEBUG oslo.service.loopingcall [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.092046] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 939.092349] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eea7da1e-0b93-4399-910a-90d4631295e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.116157] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.116157] env[61964]: value = "task-1041309" [ 939.116157] env[61964]: _type = "Task" [ 939.116157] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.126188] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041309, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.181088] env[61964]: DEBUG nova.compute.manager [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Received event network-changed-d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 939.181290] env[61964]: DEBUG nova.compute.manager [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Refreshing instance network info cache due to event network-changed-d6798518-eaec-483c-9a59-bcca91699812. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 939.181595] env[61964]: DEBUG oslo_concurrency.lockutils [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] Acquiring lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.181817] env[61964]: DEBUG oslo_concurrency.lockutils [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] Acquired lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.182009] env[61964]: DEBUG nova.network.neutron [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Refreshing network info cache for port d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.231781] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043793} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.232604] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.232868] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.233203] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk to [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 939.233519] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e4d6c8d-50e2-4529-a7ca-999138ebc88e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.240740] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 939.240740] env[61964]: value = "task-1041310" [ 939.240740] env[61964]: _type = "Task" [ 939.240740] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.250470] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.406643] env[61964]: DEBUG nova.scheduler.client.report [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 939.628404] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041309, 'name': CreateVM_Task, 'duration_secs': 0.494963} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.628629] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.629393] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.629603] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.629992] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.630291] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4496ad22-587a-4fd4-9a13-67d500797fa3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.637261] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 939.637261] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5262392f-929c-4f63-11cd-e0f5e7c8cda0" [ 939.637261] env[61964]: _type = "Task" [ 939.637261] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.647452] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5262392f-929c-4f63-11cd-e0f5e7c8cda0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.752430] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.846298] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "f5acabe4-674b-4c4c-85b4-b9270926017b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.846566] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.846791] env[61964]: DEBUG nova.compute.manager [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 939.847854] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed9dada-b3b1-44ae-b4b9-37d4b672edc1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.855909] env[61964]: DEBUG nova.compute.manager [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 939.856598] env[61964]: DEBUG nova.objects.instance [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'flavor' on Instance uuid f5acabe4-674b-4c4c-85b4-b9270926017b {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.904767] env[61964]: DEBUG nova.network.neutron [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Updated VIF entry in instance network info cache for port d6798518-eaec-483c-9a59-bcca91699812. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 939.905174] env[61964]: DEBUG nova.network.neutron [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Updating instance_info_cache with network_info: [{"id": "d6798518-eaec-483c-9a59-bcca91699812", "address": "fa:16:3e:c4:4e:8d", "network": {"id": "68d4d727-79b9-4d39-a8e0-fa481cfa70c5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1233288901-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1658dcfca064b058aaac985b9c7a5de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "96d8be6c-b557-4b40-b0f5-838c62a3c904", "external-id": "nsx-vlan-transportzone-144", "segmentation_id": 144, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6798518-ea", "ovs_interfaceid": "d6798518-eaec-483c-9a59-bcca91699812", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.912231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.914842] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.188s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.916736] env[61964]: INFO nova.compute.claims [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.947425] env[61964]: INFO nova.scheduler.client.report [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted allocations for instance 8bf586a2-0b56-4f5a-9f33-d537fba00d32 [ 940.028990] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 940.056873] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 940.057159] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.057326] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 940.057514] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.057668] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 940.057818] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 940.058043] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 940.058216] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 940.058402] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 940.058574] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 940.058751] env[61964]: DEBUG nova.virt.hardware [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 940.059716] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f95d351-3db0-4053-9ce5-c07aca991c99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.069670] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66b3eb2-4ed9-4146-bad3-d8d0f37ab5b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.149750] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5262392f-929c-4f63-11cd-e0f5e7c8cda0, 'name': SearchDatastore_Task, 'duration_secs': 0.060077} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.150166] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.150513] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.150803] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.150989] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.151222] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.151538] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8929c09-748f-428c-bd4a-85d2505265da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.169377] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.170037] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 940.171641] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3168ffce-c4c8-4333-9777-21ccdea3b66f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.179891] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 940.179891] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52eddf7d-f233-caa8-ebf2-c7557b05c552" [ 940.179891] env[61964]: _type = "Task" [ 940.179891] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.189403] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52eddf7d-f233-caa8-ebf2-c7557b05c552, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.253633] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.402095] env[61964]: DEBUG nova.compute.manager [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Received event network-vif-plugged-f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 940.402386] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] Acquiring lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.402501] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] Lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.402690] env[61964]: DEBUG oslo_concurrency.lockutils [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] Lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.402938] env[61964]: DEBUG nova.compute.manager [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] No waiting events found dispatching network-vif-plugged-f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.403201] env[61964]: WARNING nova.compute.manager [req-3e2941c3-0b69-408e-8da1-10bb71b4e797 req-49d1ae43-4785-4ade-b72a-9038c62bc0da service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Received unexpected event network-vif-plugged-f53954aa-88ff-4bd2-b4d5-139af04024c5 for instance with vm_state building and task_state spawning. [ 940.410637] env[61964]: DEBUG oslo_concurrency.lockutils [req-307ab9e8-2afa-41d2-8840-ec8c08a3eb24 req-de2813de-9793-4425-8ce1-227d34f27463 service nova] Releasing lock "refresh_cache-34e00bc8-8124-4ee9-a0a4-c70bf611996c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.459144] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bedc21da-619c-4f7e-abfa-1948afc7a651 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "8bf586a2-0b56-4f5a-9f33-d537fba00d32" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.031s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.519382] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Successfully updated port: f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.692621] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52eddf7d-f233-caa8-ebf2-c7557b05c552, 'name': SearchDatastore_Task, 'duration_secs': 0.0789} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.693628] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52008428-6b66-41ce-9d31-9b52ff086667 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.701123] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 940.701123] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528aa137-e1ea-d55a-11b6-be8904376e60" [ 940.701123] env[61964]: _type = "Task" [ 940.701123] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.711372] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528aa137-e1ea-d55a-11b6-be8904376e60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.754918] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.864632] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.865055] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59697def-cb36-44a6-b485-ddef328fca19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.873935] env[61964]: DEBUG oslo_vmware.api [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 940.873935] env[61964]: value = "task-1041311" [ 940.873935] env[61964]: _type = "Task" [ 940.873935] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.884517] env[61964]: DEBUG oslo_vmware.api [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.022299] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.022299] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.022299] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.214740] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528aa137-e1ea-d55a-11b6-be8904376e60, 'name': SearchDatastore_Task, 'duration_secs': 0.078267} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.216070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.216359] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 34e00bc8-8124-4ee9-a0a4-c70bf611996c/34e00bc8-8124-4ee9-a0a4-c70bf611996c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 941.217152] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec3d4d2-de6c-446e-93f7-1f292ba1517a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.219751] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5367aec4-e8e8-4542-b9db-9020480839b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.230113] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62ec9f6-49c2-44a5-b7c5-a58749198511 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.233710] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 941.233710] env[61964]: value = "task-1041312" [ 941.233710] env[61964]: _type = "Task" [ 941.233710] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.267897] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f72ad6-f582-488a-97f5-efbc0d07e881 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.273901] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.280421] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.284131] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3d3c7c-09a6-450b-9ee7-0f66535bf3c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.300306] env[61964]: DEBUG nova.compute.provider_tree [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.385310] env[61964]: DEBUG oslo_vmware.api [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041311, 'name': PowerOffVM_Task, 'duration_secs': 0.389299} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.385635] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.385849] env[61964]: DEBUG nova.compute.manager [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.386730] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecc8e5a-664f-420e-b98f-5e74474a112e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.557584] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.744661] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041312, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.774329] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041310, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.51773} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.775741] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f732a7b4-4004-478c-ad84-b0f50659c288/f732a7b4-4004-478c-ad84-b0f50659c288.vmdk to [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 941.775741] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f5986a-5e52-4daf-8ecd-67f869d2dc9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.799280] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.799606] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92cc8517-0056-472e-935c-a78211903957 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.815643] env[61964]: DEBUG nova.scheduler.client.report [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.825726] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 941.825726] env[61964]: value = "task-1041313" [ 941.825726] env[61964]: _type = "Task" [ 941.825726] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.833272] env[61964]: DEBUG nova.network.neutron [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Updating instance_info_cache with network_info: [{"id": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "address": "fa:16:3e:88:9c:76", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf53954aa-88", "ovs_interfaceid": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.847838] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041313, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.900268] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a5ba9401-a551-44f8-b415-62114b3f270f tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.245093] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041312, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.323580] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.324136] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 942.326784] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.360s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.327045] env[61964]: DEBUG nova.objects.instance [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lazy-loading 'resources' on Instance uuid ba0f227f-640a-4765-a84e-61bd81128356 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.338262] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.338561] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Instance network_info: |[{"id": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "address": "fa:16:3e:88:9c:76", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf53954aa-88", "ovs_interfaceid": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 942.338854] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041313, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.339234] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:9c:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f53954aa-88ff-4bd2-b4d5-139af04024c5', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.346798] env[61964]: DEBUG oslo.service.loopingcall [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.347281] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.347522] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98cbbc0c-4fed-43ae-be79-c806d3a1c35a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.370637] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.370637] env[61964]: value = "task-1041314" [ 942.370637] env[61964]: _type = "Task" [ 942.370637] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.379191] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041314, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.431264] env[61964]: DEBUG nova.compute.manager [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Received event network-changed-f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 942.431542] env[61964]: DEBUG nova.compute.manager [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Refreshing instance network info cache due to event network-changed-f53954aa-88ff-4bd2-b4d5-139af04024c5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 942.431823] env[61964]: DEBUG oslo_concurrency.lockutils [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] Acquiring lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.431972] env[61964]: DEBUG oslo_concurrency.lockutils [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] Acquired lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.432156] env[61964]: DEBUG nova.network.neutron [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Refreshing network info cache for port f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.535465] env[61964]: DEBUG nova.objects.instance [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'flavor' on Instance uuid f5acabe4-674b-4c4c-85b4-b9270926017b {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.746203] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041312, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.13039} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.746470] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 34e00bc8-8124-4ee9-a0a4-c70bf611996c/34e00bc8-8124-4ee9-a0a4-c70bf611996c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 942.746684] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.746942] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee3a4afd-a1e6-45b1-b60e-02a52ebcd7b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.754729] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 942.754729] env[61964]: value = "task-1041315" [ 942.754729] env[61964]: _type = "Task" [ 942.754729] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.763309] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.829835] env[61964]: DEBUG nova.compute.utils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.836409] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 942.836776] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.843902] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 942.856055] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041313, 'name': ReconfigVM_Task, 'duration_secs': 1.001615} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.856055] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449/63793ce6-d511-403e-8a4b-cad8c4157449.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.856055] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17c75b7b-eaec-4ab5-b6ca-7faaed3b8b91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.864556] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 942.864556] env[61964]: value = "task-1041316" [ 942.864556] env[61964]: _type = "Task" [ 942.864556] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.881184] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041316, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.888231] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041314, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.892723] env[61964]: DEBUG nova.policy [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.041778] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.041965] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquired lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.042167] env[61964]: DEBUG nova.network.neutron [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.042352] env[61964]: DEBUG nova.objects.instance [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'info_cache' on Instance uuid f5acabe4-674b-4c4c-85b4-b9270926017b {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.140245] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2a16e9-f7fc-4af5-a908-0fa766d92359 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.149637] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167ea7fb-c406-4646-83b5-295e9038480b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.184254] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65c4fdc-0338-44b7-aaa3-2aa8d1452896 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.194236] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8749c7-f335-488e-a518-6d23090114b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.201905] env[61964]: DEBUG nova.network.neutron [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Updated VIF entry in instance network info cache for port f53954aa-88ff-4bd2-b4d5-139af04024c5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.202647] env[61964]: DEBUG nova.network.neutron [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Updating instance_info_cache with network_info: [{"id": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "address": "fa:16:3e:88:9c:76", "network": {"id": "88f67ca5-e62c-49ba-a425-26683d6bd97f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1075310591-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3f1804a0ded74317a1e2016db18c55ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf53954aa-88", "ovs_interfaceid": "f53954aa-88ff-4bd2-b4d5-139af04024c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.215219] env[61964]: DEBUG nova.compute.provider_tree [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.266418] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112956} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.266418] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.267191] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b950bee-d8b7-45cc-8c45-a29699414a37 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.273057] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Successfully created port: 6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.296928] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 34e00bc8-8124-4ee9-a0a4-c70bf611996c/34e00bc8-8124-4ee9-a0a4-c70bf611996c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.297703] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbe37f26-501a-489c-be57-36cd32463b0e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.319132] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 943.319132] env[61964]: value = "task-1041317" [ 943.319132] env[61964]: _type = "Task" [ 943.319132] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.328322] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041317, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.375172] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041316, 'name': Rename_Task, 'duration_secs': 0.204084} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.378343] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 943.378608] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63a8bc34-2631-405e-8ff4-d40cb8033a53 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.387195] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041314, 'name': CreateVM_Task, 'duration_secs': 0.597944} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.388459] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 943.388801] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 943.388801] env[61964]: value = "task-1041318" [ 943.388801] env[61964]: _type = "Task" [ 943.388801] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.390054] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.390232] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.390561] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.391118] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-063a20c5-68fc-4b49-9cdb-82b1ad98660a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.401098] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 943.401098] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dd3288-eb6e-948d-57c6-88b1a3aa62e1" [ 943.401098] env[61964]: _type = "Task" [ 943.401098] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.404310] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.413263] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dd3288-eb6e-948d-57c6-88b1a3aa62e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.546462] env[61964]: DEBUG nova.objects.base [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.706228] env[61964]: DEBUG oslo_concurrency.lockutils [req-0aa909b7-4203-4e18-b7a6-345876090035 req-945d03a9-bc83-4d3d-b4c7-6ea6778f91dd service nova] Releasing lock "refresh_cache-58f1ad74-ea72-45c8-b98c-c85674989342" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.719511] env[61964]: DEBUG nova.scheduler.client.report [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 943.830647] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.854373] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 943.901171] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041318, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.910703] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 943.910982] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.911166] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 943.911357] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.911509] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 943.911660] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 943.911870] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 943.912043] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 943.912218] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 943.912381] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 943.912594] env[61964]: DEBUG nova.virt.hardware [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 943.913444] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19bab3d-00f8-4b79-9a2b-3f50eef8d5b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.919482] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dd3288-eb6e-948d-57c6-88b1a3aa62e1, 'name': SearchDatastore_Task, 'duration_secs': 0.019229} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.920135] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.920382] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.920626] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.920935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.920935] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.921264] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08600ef8-28d0-4771-ad28-d273f3c75a35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.926510] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630d63f8-da79-4a9f-9ba5-fe5f51f7341f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.932783] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.932989] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.941264] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ac06ec1-8dae-4f1e-ae68-27d160015e07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.948045] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 943.948045] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5240735f-e635-8477-32ff-c2474f8b2d3c" [ 943.948045] env[61964]: _type = "Task" [ 943.948045] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.959208] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5240735f-e635-8477-32ff-c2474f8b2d3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.224527] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.230328] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.923s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.230884] env[61964]: DEBUG nova.objects.instance [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'resources' on Instance uuid f73b99bf-6a2b-4f21-b855-74965c95ed76 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.260319] env[61964]: INFO nova.scheduler.client.report [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleted allocations for instance ba0f227f-640a-4765-a84e-61bd81128356 [ 944.270576] env[61964]: DEBUG nova.network.neutron [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updating instance_info_cache with network_info: [{"id": "c820ead7-034a-450f-839a-9bca2af7bb43", "address": "fa:16:3e:dc:4b:85", "network": {"id": "a026c320-8d76-4143-9ff1-a5873d9cbddc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-458375957-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd511575612441cc92145aa768353f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc820ead7-03", "ovs_interfaceid": "c820ead7-034a-450f-839a-9bca2af7bb43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.331848] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.401312] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041318, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.458715] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5240735f-e635-8477-32ff-c2474f8b2d3c, 'name': SearchDatastore_Task, 'duration_secs': 0.010316} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.459892] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1789a0f-1e19-4126-9f64-71f0c14f1b37 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.465223] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 944.465223] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200d7f3-718b-bf12-0f73-36d3beb0276c" [ 944.465223] env[61964]: _type = "Task" [ 944.465223] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.472950] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200d7f3-718b-bf12-0f73-36d3beb0276c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.734182] env[61964]: DEBUG nova.objects.instance [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'numa_topology' on Instance uuid f73b99bf-6a2b-4f21-b855-74965c95ed76 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.758906] env[61964]: DEBUG nova.compute.manager [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Received event network-vif-plugged-6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 944.759138] env[61964]: DEBUG oslo_concurrency.lockutils [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] Acquiring lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.759345] env[61964]: DEBUG oslo_concurrency.lockutils [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.759512] env[61964]: DEBUG oslo_concurrency.lockutils [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.759777] env[61964]: DEBUG nova.compute.manager [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] No waiting events found dispatching network-vif-plugged-6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 944.759993] env[61964]: WARNING nova.compute.manager [req-9edc4117-8cf6-49a2-b4c5-218ee7738f5b req-5649cd41-fb54-416d-bd75-320e6f3495cb service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Received unexpected event network-vif-plugged-6b04d710-a04d-43a5-839f-29a9451ae783 for instance with vm_state building and task_state spawning. [ 944.769638] env[61964]: DEBUG oslo_concurrency.lockutils [None req-e864b7b1-cff2-4d03-a50e-ee8346130915 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "ba0f227f-640a-4765-a84e-61bd81128356" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.998s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.774796] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Releasing lock "refresh_cache-f5acabe4-674b-4c4c-85b4-b9270926017b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.833746] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041317, 'name': ReconfigVM_Task, 'duration_secs': 1.06992} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.836969] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 34e00bc8-8124-4ee9-a0a4-c70bf611996c/34e00bc8-8124-4ee9-a0a4-c70bf611996c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.837626] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d135f547-51fa-45e1-b0bd-1f06f4b4abe3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.843960] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 944.843960] env[61964]: value = "task-1041319" [ 944.843960] env[61964]: _type = "Task" [ 944.843960] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.853957] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041319, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.856895] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Successfully updated port: 6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.901848] env[61964]: DEBUG oslo_vmware.api [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041318, 'name': PowerOnVM_Task, 'duration_secs': 1.111709} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.902414] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 944.928608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.928608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.929349] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "52f358e4-beb2-4b96-8f96-5cd4da103eda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.929349] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.929349] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.931762] env[61964]: INFO nova.compute.manager [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Terminating instance [ 944.975412] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200d7f3-718b-bf12-0f73-36d3beb0276c, 'name': SearchDatastore_Task, 'duration_secs': 0.013197} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.975719] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.975979] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 58f1ad74-ea72-45c8-b98c-c85674989342/58f1ad74-ea72-45c8-b98c-c85674989342.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.976256] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f592404-75bc-4d52-b44e-7a6eb5081640 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.983587] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 944.983587] env[61964]: value = "task-1041320" [ 944.983587] env[61964]: _type = "Task" [ 944.983587] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.991091] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041320, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.004255] env[61964]: DEBUG nova.compute.manager [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 945.005123] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d917b42-4c0b-446c-aecc-4a41117f72b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.236884] env[61964]: DEBUG nova.objects.base [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 945.353795] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041319, 'name': Rename_Task, 'duration_secs': 0.269551} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.356282] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.356731] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-414b5c2d-7ca1-4fd0-9b90-e27f092ab594 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.358874] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.359018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.359167] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.363969] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 945.363969] env[61964]: value = "task-1041321" [ 945.363969] env[61964]: _type = "Task" [ 945.363969] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.375411] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.438515] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "refresh_cache-52f358e4-beb2-4b96-8f96-5cd4da103eda" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.438757] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquired lock "refresh_cache-52f358e4-beb2-4b96-8f96-5cd4da103eda" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.438971] env[61964]: DEBUG nova.network.neutron [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.474039] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60754176-d75d-409b-b5a1-3d57944567fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.482418] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d8436d-b9b7-4b1f-9c3e-2f95034d9b20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.496856] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041320, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.531138] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9594de98-2ee5-4790-8e97-99d1502d53a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.534709] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95cebdc7-d048-4f6d-818c-0f2cc07af73d tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 48.269s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.541661] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4125d611-182a-4812-8f16-6a954b41c327 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.558025] env[61964]: DEBUG nova.compute.provider_tree [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.781013] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.781760] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7b517ea-4228-4e50-bd3f-51487f6a31a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.788732] env[61964]: DEBUG oslo_vmware.api [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 945.788732] env[61964]: value = "task-1041322" [ 945.788732] env[61964]: _type = "Task" [ 945.788732] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.797020] env[61964]: DEBUG oslo_vmware.api [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.873678] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041321, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.916501] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 945.958437] env[61964]: DEBUG nova.network.neutron [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 945.997124] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041320, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.765715} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.997450] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 58f1ad74-ea72-45c8-b98c-c85674989342/58f1ad74-ea72-45c8-b98c-c85674989342.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 945.997792] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.997923] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37e6a83d-179f-4f99-8be3-ca77f83f5ba5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.005907] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 946.005907] env[61964]: value = "task-1041323" [ 946.005907] env[61964]: _type = "Task" [ 946.005907] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.015084] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.053406] env[61964]: DEBUG nova.network.neutron [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.062369] env[61964]: DEBUG nova.scheduler.client.report [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.088363] env[61964]: DEBUG nova.network.neutron [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Updating instance_info_cache with network_info: [{"id": "6b04d710-a04d-43a5-839f-29a9451ae783", "address": "fa:16:3e:0d:81:48", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b04d710-a0", "ovs_interfaceid": "6b04d710-a04d-43a5-839f-29a9451ae783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.299752] env[61964]: DEBUG oslo_vmware.api [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041322, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.372987] env[61964]: DEBUG oslo_vmware.api [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041321, 'name': PowerOnVM_Task, 'duration_secs': 0.933021} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.373298] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.373505] env[61964]: INFO nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Took 9.85 seconds to spawn the instance on the hypervisor. [ 946.373688] env[61964]: DEBUG nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.374490] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c3c972-d5cf-4a74-9016-ffb473c7dacf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.515477] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062919} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.515776] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.516571] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533f2dcb-d015-4a6b-961b-5c9c03ea1e57 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.539046] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 58f1ad74-ea72-45c8-b98c-c85674989342/58f1ad74-ea72-45c8-b98c-c85674989342.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.539046] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-897bbd82-58ab-4c0c-b2e4-6bc1568afe1f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.556560] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Releasing lock "refresh_cache-52f358e4-beb2-4b96-8f96-5cd4da103eda" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.557045] env[61964]: DEBUG nova.compute.manager [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 946.557248] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 946.559085] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b819b8b-6404-4834-8e85-9244cce99d28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.562252] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 946.562252] env[61964]: value = "task-1041324" [ 946.562252] env[61964]: _type = "Task" [ 946.562252] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.567869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.338s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.570184] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 946.571043] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.624s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.572227] env[61964]: INFO nova.compute.claims [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.575708] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c581e7fd-557f-4dd5-bce1-498d1191c90d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.581174] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041324, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.582416] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 946.582416] env[61964]: value = "task-1041325" [ 946.582416] env[61964]: _type = "Task" [ 946.582416] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.590715] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.590963] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Instance network_info: |[{"id": "6b04d710-a04d-43a5-839f-29a9451ae783", "address": "fa:16:3e:0d:81:48", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b04d710-a0", "ovs_interfaceid": "6b04d710-a04d-43a5-839f-29a9451ae783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 946.591267] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.591665] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:81:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b04d710-a04d-43a5-839f-29a9451ae783', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.599483] env[61964]: DEBUG oslo.service.loopingcall [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.600777] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 946.602388] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57b5a0dd-2ade-4e4c-9a7e-d32243f07387 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.621765] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.621765] env[61964]: value = "task-1041326" [ 946.621765] env[61964]: _type = "Task" [ 946.621765] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.630692] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041326, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.802240] env[61964]: DEBUG oslo_vmware.api [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041322, 'name': PowerOnVM_Task, 'duration_secs': 0.565244} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.802538] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.802847] env[61964]: DEBUG nova.compute.manager [None req-c35ca096-6aa7-42a8-be39-c19518eb0103 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.803609] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf8438d-68fe-4522-97c7-1c7390ef932c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.809743] env[61964]: DEBUG nova.compute.manager [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Received event network-changed-6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 946.809950] env[61964]: DEBUG nova.compute.manager [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Refreshing instance network info cache due to event network-changed-6b04d710-a04d-43a5-839f-29a9451ae783. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 946.810163] env[61964]: DEBUG oslo_concurrency.lockutils [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] Acquiring lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.810305] env[61964]: DEBUG oslo_concurrency.lockutils [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] Acquired lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.810465] env[61964]: DEBUG nova.network.neutron [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Refreshing network info cache for port 6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.884846] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c273289-2ca7-45a2-be16-a346c4b6aeda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.897271] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Suspending the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 946.897941] env[61964]: INFO nova.compute.manager [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Took 41.76 seconds to build instance. [ 946.898780] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-f9fe6b9c-0cc9-44b1-b8a8-8abf79825c60 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.904936] env[61964]: DEBUG oslo_vmware.api [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 946.904936] env[61964]: value = "task-1041327" [ 946.904936] env[61964]: _type = "Task" [ 946.904936] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.914107] env[61964]: DEBUG oslo_vmware.api [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041327, 'name': SuspendVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.073834] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041324, 'name': ReconfigVM_Task, 'duration_secs': 0.425568} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.076991] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 58f1ad74-ea72-45c8-b98c-c85674989342/58f1ad74-ea72-45c8-b98c-c85674989342.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.077485] env[61964]: DEBUG oslo_concurrency.lockutils [None req-18360b86-323c-424f-b734-27c7afca846f tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 51.802s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.080073] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffaac14d-c7d7-45eb-aab7-196972d14bd4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.081903] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 29.478s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.082141] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.082809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.082809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.084575] env[61964]: INFO nova.compute.manager [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Terminating instance [ 947.093602] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 947.093602] env[61964]: value = "task-1041328" [ 947.093602] env[61964]: _type = "Task" [ 947.093602] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.100244] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041325, 'name': PowerOffVM_Task, 'duration_secs': 0.199634} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.100910] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.101136] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.101641] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9605420f-56de-45b5-8c64-f93cc5005c19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.106893] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041328, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.132815] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041326, 'name': CreateVM_Task, 'duration_secs': 0.390702} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.133653] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 947.134384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.134566] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.134970] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.135795] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-597f9bb3-cd5e-405b-b0d0-b3781950ea27 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.137896] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 947.137896] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 947.138080] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleting the datastore file [datastore2] 52f358e4-beb2-4b96-8f96-5cd4da103eda {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.138672] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b154ac8-c9a6-488c-a9ad-9d8d59ded863 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.144116] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 947.144116] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5244de74-893e-de08-cc48-9f600fd301bd" [ 947.144116] env[61964]: _type = "Task" [ 947.144116] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.148834] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for the task: (returnval){ [ 947.148834] env[61964]: value = "task-1041330" [ 947.148834] env[61964]: _type = "Task" [ 947.148834] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.155081] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5244de74-893e-de08-cc48-9f600fd301bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.160453] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.401591] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a699d9a9-98b1-462c-a1aa-b85341ecb120 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.268s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.417476] env[61964]: DEBUG oslo_vmware.api [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041327, 'name': SuspendVM_Task} progress is 70%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.540680] env[61964]: DEBUG nova.network.neutron [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Updated VIF entry in instance network info cache for port 6b04d710-a04d-43a5-839f-29a9451ae783. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 947.540680] env[61964]: DEBUG nova.network.neutron [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Updating instance_info_cache with network_info: [{"id": "6b04d710-a04d-43a5-839f-29a9451ae783", "address": "fa:16:3e:0d:81:48", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b04d710-a0", "ovs_interfaceid": "6b04d710-a04d-43a5-839f-29a9451ae783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.595257] env[61964]: DEBUG nova.compute.manager [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 947.595525] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.595880] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd58db3c-09da-4081-80df-b19a4939b64e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.608189] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041328, 'name': Rename_Task, 'duration_secs': 0.212789} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.609691] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 947.609788] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b705876-df99-4264-be58-2fb3f6fa5e6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.614230] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef75e08-9aae-42e3-81d6-4d3669812597 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.633047] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 947.633047] env[61964]: value = "task-1041331" [ 947.633047] env[61964]: _type = "Task" [ 947.633047] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.652312] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041331, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.652918] env[61964]: WARNING nova.virt.vmwareapi.vmops [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f73b99bf-6a2b-4f21-b855-74965c95ed76 could not be found. [ 947.653131] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.653317] env[61964]: INFO nova.compute.manager [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Took 0.06 seconds to destroy the instance on the hypervisor. [ 947.653557] env[61964]: DEBUG oslo.service.loopingcall [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.661773] env[61964]: DEBUG nova.compute.manager [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 947.661886] env[61964]: DEBUG nova.network.neutron [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.669477] env[61964]: DEBUG oslo_vmware.api [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Task: {'id': task-1041330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.395888} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.673332] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.673545] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 947.673738] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.673976] env[61964]: INFO nova.compute.manager [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Took 1.12 seconds to destroy the instance on the hypervisor. [ 947.674254] env[61964]: DEBUG oslo.service.loopingcall [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.674472] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5244de74-893e-de08-cc48-9f600fd301bd, 'name': SearchDatastore_Task, 'duration_secs': 0.047059} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.677251] env[61964]: DEBUG nova.compute.manager [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 947.677345] env[61964]: DEBUG nova.network.neutron [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 947.679027] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.679233] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.679891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.679891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.679891] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.681132] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97753eed-5487-4bcd-ac94-f6f6cdafdcd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.692054] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.692289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 947.693078] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ac948ae-8e0f-4da7-b7b3-af7ce7278de8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.702383] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 947.702383] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520686b5-df2d-5ae1-1703-d01e4777baed" [ 947.702383] env[61964]: _type = "Task" [ 947.702383] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.707070] env[61964]: DEBUG nova.network.neutron [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 947.713639] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520686b5-df2d-5ae1-1703-d01e4777baed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.884843] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d49f6cf-80ee-4f1e-945b-caf8a8ae41b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.894670] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e92ac8-c587-4776-8fa2-645aa08d11bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.932910] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5559f6-2de4-4fc7-95c4-07ac64b91bcc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.941892] env[61964]: DEBUG oslo_vmware.api [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041327, 'name': SuspendVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.946617] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b27e6c-44f7-4be9-a65f-9399524e0d10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.962085] env[61964]: DEBUG nova.compute.provider_tree [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.043722] env[61964]: DEBUG oslo_concurrency.lockutils [req-633fe644-6245-4ad7-88b8-1de75a351999 req-a4626da2-2913-4db9-aa22-ee03f5cf4b51 service nova] Releasing lock "refresh_cache-2c77a07d-c949-4aef-8d6f-132969d58cee" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.144867] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041331, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.213870] env[61964]: DEBUG nova.network.neutron [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.215934] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520686b5-df2d-5ae1-1703-d01e4777baed, 'name': SearchDatastore_Task, 'duration_secs': 0.022127} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.219118] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dbe9c0a-4623-49dc-be01-0a35df95d35f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.230136] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 948.230136] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525bbc11-5f32-ee83-50b2-4642d140e8e8" [ 948.230136] env[61964]: _type = "Task" [ 948.230136] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.240413] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525bbc11-5f32-ee83-50b2-4642d140e8e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.446020] env[61964]: DEBUG oslo_vmware.api [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041327, 'name': SuspendVM_Task, 'duration_secs': 1.047658} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.446378] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Suspended the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 948.446636] env[61964]: DEBUG nova.compute.manager [None req-bdef8e8a-347a-4f62-a02b-5b959e278af6 tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.447448] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f941cb-00f9-4ad7-8f18-1cdf17d358a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.465458] env[61964]: DEBUG nova.scheduler.client.report [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.478232] env[61964]: DEBUG nova.network.neutron [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.644275] env[61964]: DEBUG oslo_vmware.api [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041331, 'name': PowerOnVM_Task, 'duration_secs': 0.65459} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.644421] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 948.644670] env[61964]: INFO nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Took 8.62 seconds to spawn the instance on the hypervisor. [ 948.644948] env[61964]: DEBUG nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 948.646027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96a2ff3-a509-46a6-b1e9-dbc4eec6d2e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.717455] env[61964]: INFO nova.compute.manager [-] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Took 1.04 seconds to deallocate network for instance. [ 948.741276] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525bbc11-5f32-ee83-50b2-4642d140e8e8, 'name': SearchDatastore_Task, 'duration_secs': 0.023257} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.741602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.741823] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 2c77a07d-c949-4aef-8d6f-132969d58cee/2c77a07d-c949-4aef-8d6f-132969d58cee.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.742100] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9828496b-3798-4d8a-883c-b8ce7bc7b33f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.748624] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 948.748624] env[61964]: value = "task-1041332" [ 948.748624] env[61964]: _type = "Task" [ 948.748624] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.757986] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041332, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.971790] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.972490] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 948.977618] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.927s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.977859] env[61964]: DEBUG nova.objects.instance [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'resources' on Instance uuid df0a9a69-bd44-4da4-ba3a-9ba241c010a4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.986023] env[61964]: INFO nova.compute.manager [-] [instance: f73b99bf-6a2b-4f21-b855-74965c95ed76] Took 1.32 seconds to deallocate network for instance. [ 949.165297] env[61964]: INFO nova.compute.manager [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Took 37.09 seconds to build instance. [ 949.225084] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.263337] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041332, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.351420] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.351682] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.351938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.352167] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.352355] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.354995] env[61964]: INFO nova.compute.manager [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Terminating instance [ 949.482573] env[61964]: DEBUG nova.compute.utils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.487185] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 949.487294] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.541518] env[61964]: DEBUG nova.policy [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.605312] env[61964]: INFO nova.compute.manager [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Resuming [ 949.605893] env[61964]: DEBUG nova.objects.instance [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'flavor' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.667261] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f4e3f40-4996-4f3e-9ed8-f7608e5e14db tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.604s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.732288] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.732562] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.732777] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.732970] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.733163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.736168] env[61964]: INFO nova.compute.manager [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Terminating instance [ 949.748147] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b88f6c-c028-4b54-91ce-71f24f9a4332 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.769091] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041332, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527314} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.770890] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc893009-0783-46c4-8c8a-ed27ec51102d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.774641] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 2c77a07d-c949-4aef-8d6f-132969d58cee/2c77a07d-c949-4aef-8d6f-132969d58cee.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 949.775225] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.775644] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6900abb8-7c15-4541-8c56-aff77faaf14b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.808030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa1bcd9-b124-40cb-97f9-4dbb08846f8b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.810113] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 949.810113] env[61964]: value = "task-1041333" [ 949.810113] env[61964]: _type = "Task" [ 949.810113] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.816740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48df4a34-5195-4a40-9478-13520b9ced99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.824163] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041333, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.833964] env[61964]: DEBUG nova.compute.provider_tree [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.858820] env[61964]: DEBUG nova.compute.manager [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 949.859059] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.859910] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218a5820-b8a9-40e8-a74c-44e1bdb4d4c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.868969] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.869229] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee7345a3-d258-49be-8b2d-da536fe2e25d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.875897] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 949.875897] env[61964]: value = "task-1041334" [ 949.875897] env[61964]: _type = "Task" [ 949.875897] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.885078] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.928248] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Successfully created port: bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.987852] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 950.016704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-107404d9-51ab-4c28-8b3b-b5cad9a265bb tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f73b99bf-6a2b-4f21-b855-74965c95ed76" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.934s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.244874] env[61964]: DEBUG nova.compute.manager [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 950.244874] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.244874] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b93a24-7672-450d-8b25-68579d323e50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.252660] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.252891] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fcdd8c6-1d52-4026-b28d-60d09a911d92 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.259226] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 950.259226] env[61964]: value = "task-1041335" [ 950.259226] env[61964]: _type = "Task" [ 950.259226] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.267741] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041335, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.323610] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041333, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064552} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.323937] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.326987] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb96865-28ab-4ff4-911b-b175aaf9c015 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.341914] env[61964]: DEBUG nova.scheduler.client.report [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.353405] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 2c77a07d-c949-4aef-8d6f-132969d58cee/2c77a07d-c949-4aef-8d6f-132969d58cee.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.353901] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3829590-beae-49f2-81f4-7089b3fa2596 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.374376] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 950.374376] env[61964]: value = "task-1041336" [ 950.374376] env[61964]: _type = "Task" [ 950.374376] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.385660] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041334, 'name': PowerOffVM_Task, 'duration_secs': 0.205199} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.388580] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.388762] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.389018] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041336, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.389243] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1d26ad7-9da6-4eb6-a5b1-3e629444c667 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.451861] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.451861] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.451861] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Deleting the datastore file [datastore2] 34e00bc8-8124-4ee9-a0a4-c70bf611996c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.451861] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c3e2839-6e96-42e2-8280-a8d275fd1683 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.458059] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for the task: (returnval){ [ 950.458059] env[61964]: value = "task-1041338" [ 950.458059] env[61964]: _type = "Task" [ 950.458059] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.465238] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.612530] env[61964]: DEBUG nova.compute.manager [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 950.613421] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6057ae85-b727-4cfc-8f07-ca801a3a372a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.769774] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041335, 'name': PowerOffVM_Task, 'duration_secs': 0.428082} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.769774] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.769774] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.770000] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02366eb4-bb74-4531-b892-cdc20732f2c2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.843229] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.843473] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.843661] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Deleting the datastore file [datastore2] f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.843932] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a9fdc6e-ec38-4398-a641-133bb2b1e670 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.850385] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for the task: (returnval){ [ 950.850385] env[61964]: value = "task-1041340" [ 950.850385] env[61964]: _type = "Task" [ 950.850385] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.855548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.860658] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.298s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.862141] env[61964]: INFO nova.compute.claims [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.864738] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.884190] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041336, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.885302] env[61964]: INFO nova.scheduler.client.report [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted allocations for instance df0a9a69-bd44-4da4-ba3a-9ba241c010a4 [ 950.971831] env[61964]: DEBUG oslo_vmware.api [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Task: {'id': task-1041338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126864} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.972356] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.972730] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.973139] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.973505] env[61964]: INFO nova.compute.manager [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 950.973947] env[61964]: DEBUG oslo.service.loopingcall [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.974316] env[61964]: DEBUG nova.compute.manager [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 950.974525] env[61964]: DEBUG nova.network.neutron [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.998853] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 951.025435] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 951.025844] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.026138] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 951.026463] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.026725] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 951.026986] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 951.027359] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 951.027686] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 951.028012] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 951.028311] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 951.028611] env[61964]: DEBUG nova.virt.hardware [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 951.029209] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "d8f24177-b695-4d40-958d-9272b4f683cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.029564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.029892] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.030224] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.030520] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.033665] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653eb8e7-9c07-449c-9e3f-6a57cc054ad6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.038706] env[61964]: INFO nova.compute.manager [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Terminating instance [ 951.048411] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c85dcc-88d5-4e64-b84c-28165e3821e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.119130] env[61964]: DEBUG oslo_concurrency.lockutils [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.119384] env[61964]: DEBUG oslo_concurrency.lockutils [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquired lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.119575] env[61964]: DEBUG nova.network.neutron [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 951.125492] env[61964]: INFO nova.compute.manager [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] instance snapshotting [ 951.128030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97aab16-fd03-4d09-9c46-b844205fae13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.148521] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47da71e9-3efb-4b9f-8083-0d126cdf19d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.341252] env[61964]: DEBUG nova.compute.manager [req-55886f5c-927b-4032-a4ac-55a18b6e7ef8 req-39041f77-1402-4c6e-9560-1e4adfc5a8a9 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Received event network-vif-deleted-d6798518-eaec-483c-9a59-bcca91699812 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 951.341462] env[61964]: INFO nova.compute.manager [req-55886f5c-927b-4032-a4ac-55a18b6e7ef8 req-39041f77-1402-4c6e-9560-1e4adfc5a8a9 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Neutron deleted interface d6798518-eaec-483c-9a59-bcca91699812; detaching it from the instance and deleting it from the info cache [ 951.341640] env[61964]: DEBUG nova.network.neutron [req-55886f5c-927b-4032-a4ac-55a18b6e7ef8 req-39041f77-1402-4c6e-9560-1e4adfc5a8a9 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.362940] env[61964]: DEBUG oslo_vmware.api [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Task: {'id': task-1041340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132755} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.364277] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.364566] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.364810] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.365093] env[61964]: INFO nova.compute.manager [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 951.365396] env[61964]: DEBUG oslo.service.loopingcall [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.367997] env[61964]: DEBUG nova.compute.manager [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 951.368241] env[61964]: DEBUG nova.network.neutron [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.387044] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041336, 'name': ReconfigVM_Task, 'duration_secs': 0.606481} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.387203] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 2c77a07d-c949-4aef-8d6f-132969d58cee/2c77a07d-c949-4aef-8d6f-132969d58cee.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.388090] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26de37ff-280b-4a14-9147-3d1876074e28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.395208] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b928a0de-09af-4238-96df-0dcad9864e01 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "df0a9a69-bd44-4da4-ba3a-9ba241c010a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.567s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.400693] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 951.400693] env[61964]: value = "task-1041341" [ 951.400693] env[61964]: _type = "Task" [ 951.400693] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.413480] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041341, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.544746] env[61964]: DEBUG nova.compute.manager [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 951.545043] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 951.546805] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f30faf-1450-4bcc-a0b4-5d4f158b21fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.555301] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 951.555951] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98f27c8b-3651-455c-a767-69083e49545c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.562989] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 951.562989] env[61964]: value = "task-1041342" [ 951.562989] env[61964]: _type = "Task" [ 951.562989] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.573456] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.614964] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f983316e-e012-41fd-bf20-7d9cc5accd71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.615398] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.659030] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 951.659333] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f3edb35a-02ca-44a0-b675-dbfdc0f7f1eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.668397] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 951.668397] env[61964]: value = "task-1041343" [ 951.668397] env[61964]: _type = "Task" [ 951.668397] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.679708] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041343, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.730713] env[61964]: DEBUG nova.compute.manager [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-vif-plugged-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 951.730990] env[61964]: DEBUG oslo_concurrency.lockutils [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.731234] env[61964]: DEBUG oslo_concurrency.lockutils [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.731414] env[61964]: DEBUG oslo_concurrency.lockutils [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.731587] env[61964]: DEBUG nova.compute.manager [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] No waiting events found dispatching network-vif-plugged-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 951.731756] env[61964]: WARNING nova.compute.manager [req-b519fe12-e3bd-4faf-a600-ecf3ea5f96e5 req-918f7210-52df-4e7d-9be5-fa28b5a3a48a service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received unexpected event network-vif-plugged-bc9cb29f-849e-4bea-8532-cd8b9efcb198 for instance with vm_state building and task_state spawning. [ 951.761095] env[61964]: DEBUG nova.network.neutron [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.844745] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfc7c58a-0724-4e70-af08-d3f00101dbf3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.858196] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcedaab-e9a4-4518-806d-f53089c79e21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.875086] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Successfully updated port: bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.907274] env[61964]: DEBUG nova.compute.manager [req-55886f5c-927b-4032-a4ac-55a18b6e7ef8 req-39041f77-1402-4c6e-9560-1e4adfc5a8a9 service nova] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Detach interface failed, port_id=d6798518-eaec-483c-9a59-bcca91699812, reason: Instance 34e00bc8-8124-4ee9-a0a4-c70bf611996c could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 951.922128] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041341, 'name': Rename_Task, 'duration_secs': 0.254552} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.925082] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 951.925724] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e768baf-0587-47c6-bfd3-3afb4dac2921 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.934700] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 951.934700] env[61964]: value = "task-1041344" [ 951.934700] env[61964]: _type = "Task" [ 951.934700] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.943555] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.077283] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041342, 'name': PowerOffVM_Task, 'duration_secs': 0.321884} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.077569] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 952.080632] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 952.080632] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82747623-9290-4cd9-9605-bf3c1b13ab4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.118584] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 952.163753] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 952.165189] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 952.165189] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleting the datastore file [datastore1] d8f24177-b695-4d40-958d-9272b4f683cc {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.165189] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d4082ce-7925-4a20-8132-71a7e56d5c4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.175297] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 952.175297] env[61964]: value = "task-1041346" [ 952.175297] env[61964]: _type = "Task" [ 952.175297] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.178849] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041343, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.192474] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.216210] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7812ce7-c39b-44f6-972c-932db24a4003 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.225923] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eeef73d-def2-44b4-b432-6a86e4221192 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.263931] env[61964]: INFO nova.compute.manager [-] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Took 1.29 seconds to deallocate network for instance. [ 952.265799] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c963faf0-f267-4127-954c-b9d30e0f2236 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.271336] env[61964]: DEBUG nova.network.neutron [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [{"id": "edda620f-1259-4fbb-afa6-48aef4eda40b", "address": "fa:16:3e:b9:0b:51", "network": {"id": "c2421f0e-a5ce-4568-aec2-0a513b494d41", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-815071250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b772e2152b674da8bdc2cb4f726f6772", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "02092ea4-bae0-4e42-b0ab-abc365b4395a", "external-id": "nsx-vlan-transportzone-718", "segmentation_id": 718, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedda620f-12", "ovs_interfaceid": "edda620f-1259-4fbb-afa6-48aef4eda40b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.279321] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce46263-d47b-463a-a871-cc6267f981e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.296142] env[61964]: DEBUG nova.compute.provider_tree [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.370381] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.370773] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.382289] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.382434] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.382570] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.449630] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041344, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.518375] env[61964]: DEBUG nova.network.neutron [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.638092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.680365] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041343, 'name': CreateSnapshot_Task, 'duration_secs': 0.841235} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.683420] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 952.684206] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c53030a-5beb-4ce5-9dd7-a652d8752934 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.691355] env[61964]: DEBUG oslo_vmware.api [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150413} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.694475] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.694675] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 952.694861] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 952.695080] env[61964]: INFO nova.compute.manager [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Took 1.15 seconds to destroy the instance on the hypervisor. [ 952.695343] env[61964]: DEBUG oslo.service.loopingcall [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.697740] env[61964]: DEBUG nova.compute.manager [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 952.697836] env[61964]: DEBUG nova.network.neutron [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 952.776727] env[61964]: DEBUG oslo_concurrency.lockutils [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Releasing lock "refresh_cache-63793ce6-d511-403e-8a4b-cad8c4157449" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.778207] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f34933-bf8e-4f58-9194-a1e0e7bd9214 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.781603] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.785782] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Resuming the VM {{(pid=61964) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 952.786103] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5e1bf79-ca39-4d3b-b22d-163bc9ef0b2d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.792577] env[61964]: DEBUG oslo_vmware.api [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 952.792577] env[61964]: value = "task-1041347" [ 952.792577] env[61964]: _type = "Task" [ 952.792577] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.801085] env[61964]: DEBUG nova.scheduler.client.report [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.805110] env[61964]: DEBUG oslo_vmware.api [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.875071] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 952.925204] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 952.950129] env[61964]: DEBUG oslo_vmware.api [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041344, 'name': PowerOnVM_Task, 'duration_secs': 0.581061} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.950575] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 952.950855] env[61964]: INFO nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Took 9.10 seconds to spawn the instance on the hypervisor. [ 952.951072] env[61964]: DEBUG nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 952.951953] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b06a80-2c78-4fb5-b5fa-ed684d756209 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.023017] env[61964]: INFO nova.compute.manager [-] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Took 1.65 seconds to deallocate network for instance. [ 953.112126] env[61964]: DEBUG nova.network.neutron [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.207842] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 953.208185] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5587ac0e-de80-403c-a47b-a2bfd0f745ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.217482] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 953.217482] env[61964]: value = "task-1041348" [ 953.217482] env[61964]: _type = "Task" [ 953.217482] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.225903] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.307022] env[61964]: DEBUG oslo_vmware.api [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041347, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.307022] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.307022] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 953.309285] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.346s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.309521] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.311847] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.087s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.312522] env[61964]: DEBUG nova.objects.instance [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lazy-loading 'resources' on Instance uuid 52f358e4-beb2-4b96-8f96-5cd4da103eda {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.345305] env[61964]: INFO nova.scheduler.client.report [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocations for instance 84821045-27b2-4a99-87f9-988b8615b83f [ 953.376448] env[61964]: DEBUG nova.compute.manager [req-516446c4-a4d9-47cf-8e71-90510d926f47 req-f644e7bd-095e-498d-8769-09352219e591 service nova] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Received event network-vif-deleted-3e133918-bc4f-4a52-ac44-586ec8a513ed {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 953.394809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.428751] env[61964]: DEBUG nova.network.neutron [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.470837] env[61964]: INFO nova.compute.manager [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Took 37.76 seconds to build instance. [ 953.530662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.615354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.615754] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Instance network_info: |[{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 953.616420] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:92:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60bdba1a-14cf-46b2-9d8b-aeaf4d80c815', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc9cb29f-849e-4bea-8532-cd8b9efcb198', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.624573] env[61964]: DEBUG oslo.service.loopingcall [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.625297] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 953.625547] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26f0415e-a6ee-40b0-85d1-6e9bd63f9c5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.647351] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.647351] env[61964]: value = "task-1041349" [ 953.647351] env[61964]: _type = "Task" [ 953.647351] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.656840] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.735804] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.769223] env[61964]: DEBUG nova.compute.manager [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 953.769373] env[61964]: DEBUG nova.compute.manager [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing instance network info cache due to event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 953.769731] env[61964]: DEBUG oslo_concurrency.lockutils [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.769980] env[61964]: DEBUG oslo_concurrency.lockutils [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.770282] env[61964]: DEBUG nova.network.neutron [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 953.807094] env[61964]: DEBUG oslo_vmware.api [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041347, 'name': PowerOnVM_Task, 'duration_secs': 0.968187} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.807306] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Resumed the VM {{(pid=61964) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 953.807437] env[61964]: DEBUG nova.compute.manager [None req-165af6e1-c6fe-4ece-ac8b-d816b924333b tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 953.808317] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd377ca8-a443-4aae-a219-bd234a3ac680 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.816062] env[61964]: DEBUG nova.compute.utils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.823205] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 953.823205] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.857019] env[61964]: DEBUG oslo_concurrency.lockutils [None req-40012a70-0e68-413e-b528-0c16c56f1e04 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "84821045-27b2-4a99-87f9-988b8615b83f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.500s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.886046] env[61964]: DEBUG nova.policy [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f438e739578a4fabad898ebd5e2925cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ccf1fe5003e484fa73ce5c32767479f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.931595] env[61964]: INFO nova.compute.manager [-] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Took 1.23 seconds to deallocate network for instance. [ 953.972990] env[61964]: DEBUG oslo_concurrency.lockutils [None req-806ebe15-4fb6-4201-831c-94d876d89d2f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.284s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.107929] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1303e994-2234-45c9-aa65-d3a1bcd884f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.116802] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebae9a1-7e32-4fa2-bf75-1ed727d932e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.153662] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da689aa-ae2a-484c-a009-2ded0ec836ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.165297] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5b13fd-1cd0-4a27-8961-00600dd4dbf7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.169373] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.180507] env[61964]: DEBUG nova.compute.provider_tree [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.236581] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.323937] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 954.444173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.492465] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Successfully created port: 79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.642163] env[61964]: DEBUG nova.network.neutron [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updated VIF entry in instance network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 954.642612] env[61964]: DEBUG nova.network.neutron [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.661419] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.683407] env[61964]: DEBUG nova.scheduler.client.report [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.730044] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.145775] env[61964]: DEBUG oslo_concurrency.lockutils [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.146124] env[61964]: DEBUG nova.compute.manager [req-564d7e5e-6c88-4c73-86b7-258a3bf1507c req-6d1bbaca-e7a7-4da9-b1c4-e1e41a6474a9 service nova] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Received event network-vif-deleted-9420c5ba-2284-4fb8-a95e-74558e7d7a9b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 955.160186] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.188362] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.190714] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.553s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.192225] env[61964]: INFO nova.compute.claims [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.213817] env[61964]: INFO nova.scheduler.client.report [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Deleted allocations for instance 52f358e4-beb2-4b96-8f96-5cd4da103eda [ 955.232652] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.339943] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 955.354026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "2c77a07d-c949-4aef-8d6f-132969d58cee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.354026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.354026] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.354419] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.354419] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.356744] env[61964]: INFO nova.compute.manager [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Terminating instance [ 955.369372] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 955.369629] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.369795] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 955.369988] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.370152] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 955.370308] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 955.370523] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 955.370691] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 955.370865] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 955.371043] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 955.371228] env[61964]: DEBUG nova.virt.hardware [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 955.372910] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dfe591-1d59-4778-af96-39042d987986 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.382609] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ad83ba-21ed-4b67-8930-adba218bd88d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.663227] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.725893] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57655b7b-6e69-4683-898b-d58d732d1208 tempest-ServerShowV247Test-1375801538 tempest-ServerShowV247Test-1375801538-project-member] Lock "52f358e4-beb2-4b96-8f96-5cd4da103eda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.797s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.733044] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.861051] env[61964]: DEBUG nova.compute.manager [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 955.861330] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.862430] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db78d990-b01c-47fc-be38-5f3534627c14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.870845] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 955.871110] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa159053-4f69-4e35-9e95-50a665adde0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.877117] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 955.877117] env[61964]: value = "task-1041350" [ 955.877117] env[61964]: _type = "Task" [ 955.877117] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.888155] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.162921] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041349, 'name': CreateVM_Task, 'duration_secs': 2.348712} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.163315] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 956.163874] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.164088] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.164459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.164737] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74d8dda2-a24b-43ad-8346-c1813157243a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.170209] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 956.170209] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b203b4-1fa0-a539-913b-bc684995c28d" [ 956.170209] env[61964]: _type = "Task" [ 956.170209] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.179441] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b203b4-1fa0-a539-913b-bc684995c28d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.241539] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041348, 'name': CloneVM_Task, 'duration_secs': 2.870749} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.241863] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Created linked-clone VM from snapshot [ 956.244253] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c98bae5-b08c-4aff-8c40-1a8be7681bd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.251202] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Uploading image 385fcb16-2e13-4057-aba9-33abfec09af3 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 956.275283] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 956.276305] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d98dc8d0-d221-46de-b79a-8cd12559d346 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.284551] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 956.284551] env[61964]: value = "task-1041351" [ 956.284551] env[61964]: _type = "Task" [ 956.284551] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.297294] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041351, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.390391] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041350, 'name': PowerOffVM_Task, 'duration_secs': 0.278534} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.390391] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 956.390575] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 956.392800] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-718237f9-d11a-4cc1-85ad-00c6f3ec28d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.396590] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Successfully updated port: 79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.435148] env[61964]: DEBUG nova.compute.manager [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Received event network-vif-plugged-79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 956.435425] env[61964]: DEBUG oslo_concurrency.lockutils [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.435650] env[61964]: DEBUG oslo_concurrency.lockutils [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.436123] env[61964]: DEBUG oslo_concurrency.lockutils [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.436123] env[61964]: DEBUG nova.compute.manager [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] No waiting events found dispatching network-vif-plugged-79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 956.436283] env[61964]: WARNING nova.compute.manager [req-e73718ae-9b2e-4462-b84a-fbf6510b4e97 req-64673ea6-b7cf-4784-94aa-e10d49ec0a7f service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Received unexpected event network-vif-plugged-79400d4a-e480-4d60-8c9f-dcb1b899ae7c for instance with vm_state building and task_state spawning. [ 956.454417] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 956.454646] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 956.454914] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore2] 2c77a07d-c949-4aef-8d6f-132969d58cee {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.455802] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92343d14-07e1-470d-a950-e9704174726d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.462478] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 956.462478] env[61964]: value = "task-1041353" [ 956.462478] env[61964]: _type = "Task" [ 956.462478] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.473373] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.530099] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566147bb-1e18-4d7e-910c-e3a1930222f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.536830] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcc3adb-6df9-4231-a31b-078d1ce69418 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.569760] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70884a22-76a5-401f-8747-32f4bf443b5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.578059] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fa7b8c-6906-4dd2-ab23-a5741b9797a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.592277] env[61964]: DEBUG nova.compute.provider_tree [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.663140] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.663344] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.681223] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b203b4-1fa0-a539-913b-bc684995c28d, 'name': SearchDatastore_Task, 'duration_secs': 0.013062} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.681382] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.681697] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 956.681940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.682115] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.682305] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 956.682567] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85a1719d-dd36-4023-bd0e-57b8430097c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.690980] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 956.691201] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 956.691917] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6c696fb-83d9-4417-a6d7-ba943c0f179e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.697674] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 956.697674] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a59760-5758-59dd-1f2c-860c734ddbf4" [ 956.697674] env[61964]: _type = "Task" [ 956.697674] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.706223] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a59760-5758-59dd-1f2c-860c734ddbf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.794908] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041351, 'name': Destroy_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.899761] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.899923] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.900103] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.972867] env[61964]: DEBUG oslo_vmware.api [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200631} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.973170] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.973373] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 956.973556] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 956.973735] env[61964]: INFO nova.compute.manager [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Took 1.11 seconds to destroy the instance on the hypervisor. [ 956.974024] env[61964]: DEBUG oslo.service.loopingcall [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.974575] env[61964]: DEBUG nova.compute.manager [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 956.974693] env[61964]: DEBUG nova.network.neutron [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 957.096085] env[61964]: DEBUG nova.scheduler.client.report [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.167438] env[61964]: DEBUG nova.compute.utils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.210436] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a59760-5758-59dd-1f2c-860c734ddbf4, 'name': SearchDatastore_Task, 'duration_secs': 0.012584} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.211765] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0202bde7-11cf-4246-ba2e-ebcc396ff887 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.218541] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 957.218541] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527da142-aaf6-5966-7ec2-f8f0a8f8b075" [ 957.218541] env[61964]: _type = "Task" [ 957.218541] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.226274] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527da142-aaf6-5966-7ec2-f8f0a8f8b075, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.295093] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041351, 'name': Destroy_Task, 'duration_secs': 0.531917} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.295526] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Destroyed the VM [ 957.295834] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 957.296120] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5c70da4f-e0b6-478e-934d-6fb0efcbf732 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.302195] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 957.302195] env[61964]: value = "task-1041354" [ 957.302195] env[61964]: _type = "Task" [ 957.302195] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.310619] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041354, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.431914] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.574463] env[61964]: DEBUG nova.network.neutron [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.601054] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.601803] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 957.604942] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.823s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.605296] env[61964]: DEBUG nova.objects.instance [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lazy-loading 'resources' on Instance uuid 34e00bc8-8124-4ee9-a0a4-c70bf611996c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.671257] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.704939] env[61964]: DEBUG nova.network.neutron [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.728887] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527da142-aaf6-5966-7ec2-f8f0a8f8b075, 'name': SearchDatastore_Task, 'duration_secs': 0.017181} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.729252] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.729555] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] c4fabe85-1b95-4981-9e05-dbfedbbe1e1e/c4fabe85-1b95-4981-9e05-dbfedbbe1e1e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 957.729857] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86aa3a94-c2a1-4f1f-87fe-0ce3d116f07d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.737086] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 957.737086] env[61964]: value = "task-1041355" [ 957.737086] env[61964]: _type = "Task" [ 957.737086] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.745155] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041355, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.812554] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041354, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.077962] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.078374] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Instance network_info: |[{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 958.078846] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:01:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79400d4a-e480-4d60-8c9f-dcb1b899ae7c', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.086331] env[61964]: DEBUG oslo.service.loopingcall [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.086517] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.086744] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6385629-2a2c-46d4-8354-dab07e13512c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.106177] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.106177] env[61964]: value = "task-1041356" [ 958.106177] env[61964]: _type = "Task" [ 958.106177] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.110006] env[61964]: DEBUG nova.compute.utils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.113840] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 958.114037] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.120567] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041356, 'name': CreateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.159979] env[61964]: DEBUG nova.policy [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.207611] env[61964]: INFO nova.compute.manager [-] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Took 1.23 seconds to deallocate network for instance. [ 958.248944] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041355, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.315443] env[61964]: DEBUG oslo_vmware.api [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041354, 'name': RemoveSnapshot_Task, 'duration_secs': 0.653845} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.315443] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 958.399500] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27179ddd-23a6-42f9-ae74-da41b853abfc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.408223] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6730ab17-7cfa-4be8-af1b-58767c304698 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.441526] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8855240b-9ca6-4565-a5e2-bbe9bcb86f86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.450670] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3746b6-b621-4316-b445-c07ee8998bc0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.467056] env[61964]: DEBUG nova.compute.provider_tree [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.473287] env[61964]: DEBUG nova.compute.manager [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Received event network-changed-79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 958.473535] env[61964]: DEBUG nova.compute.manager [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Refreshing instance network info cache due to event network-changed-79400d4a-e480-4d60-8c9f-dcb1b899ae7c. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 958.473713] env[61964]: DEBUG oslo_concurrency.lockutils [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] Acquiring lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.473879] env[61964]: DEBUG oslo_concurrency.lockutils [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] Acquired lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.474044] env[61964]: DEBUG nova.network.neutron [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Refreshing network info cache for port 79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.496766] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Successfully created port: a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.615196] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 958.617653] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041356, 'name': CreateVM_Task, 'duration_secs': 0.391524} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.621018] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 958.621018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.621018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.621018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.621018] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f5fc82-91ea-4ef2-9148-c98aa7afc9f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.624213] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 958.624213] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522a847e-d58c-6d61-fb0b-8dcd6ea3cbac" [ 958.624213] env[61964]: _type = "Task" [ 958.624213] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.631799] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522a847e-d58c-6d61-fb0b-8dcd6ea3cbac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.716149] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.747907] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041355, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.745309} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.748631] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] c4fabe85-1b95-4981-9e05-dbfedbbe1e1e/c4fabe85-1b95-4981-9e05-dbfedbbe1e1e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 958.748920] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.749142] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b32fc6e5-d843-4c29-bd04-9fcb8a398743 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.755524] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 958.755524] env[61964]: value = "task-1041357" [ 958.755524] env[61964]: _type = "Task" [ 958.755524] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.759526] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.759873] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.760204] env[61964]: INFO nova.compute.manager [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Attaching volume 9930fa82-359c-40a3-9658-852e13df340f to /dev/sdb [ 958.772657] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.798327] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d99e50-44cd-49d7-aef6-67111ccb5806 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.805902] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c883b82-4838-4b71-af2d-1816f4cbcc59 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.827316] env[61964]: WARNING nova.compute.manager [None req-20b92fd3-eebe-4b1a-9900-3e3eb8ba21ad tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Image not found during snapshot: nova.exception.ImageNotFound: Image 385fcb16-2e13-4057-aba9-33abfec09af3 could not be found. [ 958.828830] env[61964]: DEBUG nova.virt.block_device [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating existing volume attachment record: 00c469d0-f954-4c89-9014-a969e2cfb8b1 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 958.969310] env[61964]: DEBUG nova.scheduler.client.report [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.137328] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522a847e-d58c-6d61-fb0b-8dcd6ea3cbac, 'name': SearchDatastore_Task, 'duration_secs': 0.085362} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.139973] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.140313] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.140568] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.142021] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.142021] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.142021] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5546ad72-1b6f-43a7-873a-b7ccf2524505 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.150899] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.151142] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.151922] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46ef7c4d-ec7f-4bcd-af42-9e6aa50393e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.158251] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 959.158251] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523f5f24-f7b6-76f5-9492-10a0c60143a4" [ 959.158251] env[61964]: _type = "Task" [ 959.158251] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.167404] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523f5f24-f7b6-76f5-9492-10a0c60143a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.251517] env[61964]: DEBUG nova.network.neutron [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updated VIF entry in instance network info cache for port 79400d4a-e480-4d60-8c9f-dcb1b899ae7c. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.251913] env[61964]: DEBUG nova.network.neutron [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.267611] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066295} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.267870] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.268663] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867495ae-731a-40d0-92cb-1ed6daa771eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.291100] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] c4fabe85-1b95-4981-9e05-dbfedbbe1e1e/c4fabe85-1b95-4981-9e05-dbfedbbe1e1e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.291426] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-885ef838-4f0c-4ec4-afb5-cffedf8d41de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.311578] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 959.311578] env[61964]: value = "task-1041361" [ 959.311578] env[61964]: _type = "Task" [ 959.311578] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.320478] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "58f1ad74-ea72-45c8-b98c-c85674989342" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.321300] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.321300] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.321300] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.321486] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.323443] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041361, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.323703] env[61964]: INFO nova.compute.manager [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Terminating instance [ 959.474485] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.476908] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.082s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.478666] env[61964]: INFO nova.compute.claims [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.497996] env[61964]: INFO nova.scheduler.client.report [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Deleted allocations for instance 34e00bc8-8124-4ee9-a0a4-c70bf611996c [ 959.626569] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 959.653930] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 959.654225] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.654409] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 959.654604] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.654915] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 959.655133] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 959.655370] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 959.655562] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 959.655755] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 959.655941] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 959.656138] env[61964]: DEBUG nova.virt.hardware [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 959.657350] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe703644-0f0b-42c8-9858-db1a5e5c3f75 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.670852] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523f5f24-f7b6-76f5-9492-10a0c60143a4, 'name': SearchDatastore_Task, 'duration_secs': 0.017416} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.672063] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5bcb1f-b859-43c0-aa49-d69d9a0b0848 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.676485] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-142ba35c-78be-4e3b-abc8-97b6818a233a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.688821] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 959.688821] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52900e65-b44f-ae5d-161a-751438fbe315" [ 959.688821] env[61964]: _type = "Task" [ 959.688821] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.696820] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52900e65-b44f-ae5d-161a-751438fbe315, 'name': SearchDatastore_Task, 'duration_secs': 0.013144} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.697074] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.697348] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 959.697590] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eac25740-ae1a-40d3-8ebf-068a3adac7ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.703724] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 959.703724] env[61964]: value = "task-1041362" [ 959.703724] env[61964]: _type = "Task" [ 959.703724] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.710631] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041362, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.754893] env[61964]: DEBUG oslo_concurrency.lockutils [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] Releasing lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.755265] env[61964]: DEBUG nova.compute.manager [req-567d5b6c-9806-41a4-ae20-81c613cfec58 req-1e1d3d92-e781-4720-93ed-7f5ede12d641 service nova] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Received event network-vif-deleted-6b04d710-a04d-43a5-839f-29a9451ae783 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 959.823895] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041361, 'name': ReconfigVM_Task, 'duration_secs': 0.333805} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.824227] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfigured VM instance instance-0000005a to attach disk [datastore2] c4fabe85-1b95-4981-9e05-dbfedbbe1e1e/c4fabe85-1b95-4981-9e05-dbfedbbe1e1e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.824884] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96d6e763-abc3-4d32-b3e4-5fc0952834f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.828426] env[61964]: DEBUG nova.compute.manager [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 959.828640] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 959.829457] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8b9778-db53-48a5-94a5-b8cc49738c3a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.833107] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 959.833107] env[61964]: value = "task-1041363" [ 959.833107] env[61964]: _type = "Task" [ 959.833107] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.838743] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.839081] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f4e6f18-21e6-4b3b-be44-a6c530f02c35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.845726] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041363, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.846967] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 959.846967] env[61964]: value = "task-1041364" [ 959.846967] env[61964]: _type = "Task" [ 959.846967] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.854780] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.007866] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bc383d22-3845-4eed-9cc9-073a69ac38e8 tempest-ServerTagsTestJSON-2043180232 tempest-ServerTagsTestJSON-2043180232-project-member] Lock "34e00bc8-8124-4ee9-a0a4-c70bf611996c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.656s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.134206] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.134498] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.134721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.135034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.135253] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.137946] env[61964]: INFO nova.compute.manager [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Terminating instance [ 960.213749] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041362, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496178} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.215792] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 960.215792] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.215792] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8bd6648-974e-46b7-9d86-42237e10212a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.220672] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 960.220672] env[61964]: value = "task-1041365" [ 960.220672] env[61964]: _type = "Task" [ 960.220672] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.228622] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.249326] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Successfully updated port: a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.343913] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041363, 'name': Rename_Task, 'duration_secs': 0.196587} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.344222] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 960.344481] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4df0126f-af3f-49e0-80f7-e165d0798246 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.351606] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 960.351606] env[61964]: value = "task-1041366" [ 960.351606] env[61964]: _type = "Task" [ 960.351606] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.357903] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041364, 'name': PowerOffVM_Task, 'duration_secs': 0.227894} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.358460] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 960.358655] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.358925] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89aac8fc-698f-4cd5-b3b9-5915941d6534 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.363041] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.424979] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.425308] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.425769] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleting the datastore file [datastore2] 58f1ad74-ea72-45c8-b98c-c85674989342 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.425877] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a970a855-2953-4960-97ee-775431a0a620 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.432293] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for the task: (returnval){ [ 960.432293] env[61964]: value = "task-1041368" [ 960.432293] env[61964]: _type = "Task" [ 960.432293] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.446837] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.499696] env[61964]: DEBUG nova.compute.manager [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Received event network-vif-plugged-a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 960.499955] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Acquiring lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.500197] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.500375] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.500552] env[61964]: DEBUG nova.compute.manager [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] No waiting events found dispatching network-vif-plugged-a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 960.500725] env[61964]: WARNING nova.compute.manager [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Received unexpected event network-vif-plugged-a70cd208-c6e3-44de-9b66-dd509b9b8a1b for instance with vm_state building and task_state spawning. [ 960.500891] env[61964]: DEBUG nova.compute.manager [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Received event network-changed-a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 960.501067] env[61964]: DEBUG nova.compute.manager [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Refreshing instance network info cache due to event network-changed-a70cd208-c6e3-44de-9b66-dd509b9b8a1b. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 960.501265] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Acquiring lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.501409] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Acquired lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.501571] env[61964]: DEBUG nova.network.neutron [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Refreshing network info cache for port a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.642931] env[61964]: DEBUG nova.compute.manager [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 960.642931] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 960.644950] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62866c2e-cee2-44b5-91c5-fa16bf82e641 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.652179] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 960.654830] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d822567-f5a7-4f84-acd5-fe7a2bbc91a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.661286] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 960.661286] env[61964]: value = "task-1041369" [ 960.661286] env[61964]: _type = "Task" [ 960.661286] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.672423] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.729410] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065327} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.729713] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.731215] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3427eaae-a251-4533-9f4c-5c5c35512b72 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.734825] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05106a6c-6952-427e-aab2-9b4998655e7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.750952] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9740a475-eecc-4a29-bc99-54f5b323ce85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.763142] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.763682] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.764251] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-318003e3-f3d9-4e5f-8597-ce9e680ed87d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.812071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92585f9-88c1-4688-ad91-38c95e9a2665 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.812251] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 960.812251] env[61964]: value = "task-1041370" [ 960.812251] env[61964]: _type = "Task" [ 960.812251] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.821146] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d083f0cc-6607-4e8e-a44b-b790e81312ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.828860] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.845816] env[61964]: DEBUG nova.compute.provider_tree [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.863577] env[61964]: DEBUG oslo_vmware.api [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041366, 'name': PowerOnVM_Task, 'duration_secs': 0.508935} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.864267] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 960.864379] env[61964]: INFO nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Took 9.87 seconds to spawn the instance on the hypervisor. [ 960.864580] env[61964]: DEBUG nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 960.865992] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb250d0-a684-4a1b-a427-f807ba9044cd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.942131] env[61964]: DEBUG oslo_vmware.api [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Task: {'id': task-1041368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149204} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.942404] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.942607] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 960.942790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 960.942975] env[61964]: INFO nova.compute.manager [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Took 1.11 seconds to destroy the instance on the hypervisor. [ 960.943248] env[61964]: DEBUG oslo.service.loopingcall [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.943516] env[61964]: DEBUG nova.compute.manager [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 960.943587] env[61964]: DEBUG nova.network.neutron [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.062110] env[61964]: DEBUG nova.network.neutron [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.160235] env[61964]: DEBUG nova.network.neutron [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.171960] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041369, 'name': PowerOffVM_Task, 'duration_secs': 0.416385} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.173897] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 961.174350] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 961.174988] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8270feeb-9aae-45be-bfd9-577e548a0441 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.252926] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 961.253418] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 961.254388] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleting the datastore file [datastore2] 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.254748] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf1688a4-310f-469b-8006-05a9496194e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.262238] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for the task: (returnval){ [ 961.262238] env[61964]: value = "task-1041373" [ 961.262238] env[61964]: _type = "Task" [ 961.262238] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.271162] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041373, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.326804] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.352132] env[61964]: DEBUG nova.scheduler.client.report [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.387188] env[61964]: INFO nova.compute.manager [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Took 43.45 seconds to build instance. [ 961.662378] env[61964]: DEBUG oslo_concurrency.lockutils [req-6a6570a1-a6c0-467e-9547-4f559ad1edcd req-20f6dc57-d6c8-4776-94d0-53c35bce0f6a service nova] Releasing lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.662810] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.663081] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.772929] env[61964]: DEBUG oslo_vmware.api [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Task: {'id': task-1041373, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153048} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.773219] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.773402] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.774019] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.774019] env[61964]: INFO nova.compute.manager [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Took 1.13 seconds to destroy the instance on the hypervisor. [ 961.774019] env[61964]: DEBUG oslo.service.loopingcall [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.774290] env[61964]: DEBUG nova.compute.manager [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 961.774392] env[61964]: DEBUG nova.network.neutron [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.824501] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041370, 'name': ReconfigVM_Task, 'duration_secs': 0.678278} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.826952] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.827714] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d2e8305-178d-4f52-998d-fd4ec8833dd3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.834394] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 961.834394] env[61964]: value = "task-1041374" [ 961.834394] env[61964]: _type = "Task" [ 961.834394] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.842984] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041374, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.861562] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.862111] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 961.865336] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.335s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.868017] env[61964]: DEBUG nova.objects.instance [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lazy-loading 'resources' on Instance uuid f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.890604] env[61964]: DEBUG oslo_concurrency.lockutils [None req-575bdfee-a881-4519-96d5-e7f1565d45bf tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.967s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.969847] env[61964]: DEBUG nova.network.neutron [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.217444] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 962.351756] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041374, 'name': Rename_Task, 'duration_secs': 0.325594} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.352335] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 962.352894] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0d00cd2-300f-4d14-b3a6-a21fd038226d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.364033] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 962.364033] env[61964]: value = "task-1041375" [ 962.364033] env[61964]: _type = "Task" [ 962.364033] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.371024] env[61964]: DEBUG nova.compute.utils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.376700] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 962.377280] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 962.379343] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.477275] env[61964]: DEBUG nova.network.neutron [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Updating instance_info_cache with network_info: [{"id": "a70cd208-c6e3-44de-9b66-dd509b9b8a1b", "address": "fa:16:3e:9a:d4:3a", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70cd208-c6", "ovs_interfaceid": "a70cd208-c6e3-44de-9b66-dd509b9b8a1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.478686] env[61964]: INFO nova.compute.manager [-] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Took 1.54 seconds to deallocate network for instance. [ 962.481195] env[61964]: DEBUG nova.policy [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '430778d9f17e4579a81902608fb3501a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b627b25af0c4982a22c8c0bd28641a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.544041] env[61964]: DEBUG nova.compute.manager [req-edc7268f-059d-4014-a9a6-415528480e2d req-d81eb00c-bbce-4074-9280-b4e9d6af51cf service nova] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Received event network-vif-deleted-f53954aa-88ff-4bd2-b4d5-139af04024c5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 962.544648] env[61964]: DEBUG nova.compute.manager [req-edc7268f-059d-4014-a9a6-415528480e2d req-d81eb00c-bbce-4074-9280-b4e9d6af51cf service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Received event network-vif-deleted-edda620f-1259-4fbb-afa6-48aef4eda40b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 962.544856] env[61964]: INFO nova.compute.manager [req-edc7268f-059d-4014-a9a6-415528480e2d req-d81eb00c-bbce-4074-9280-b4e9d6af51cf service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Neutron deleted interface edda620f-1259-4fbb-afa6-48aef4eda40b; detaching it from the instance and deleting it from the info cache [ 962.545123] env[61964]: DEBUG nova.network.neutron [req-edc7268f-059d-4014-a9a6-415528480e2d req-d81eb00c-bbce-4074-9280-b4e9d6af51cf service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.668038] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898a2b3d-3980-41ae-ac23-292a8ea86f92 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.678660] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96234d4-e4eb-4ead-88b0-c5daebfe3a3e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.711959] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a3181c-cdd3-49b8-b07e-372637ff6c7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.722342] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c77146-47f1-4202-badc-27dfe53414cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.740884] env[61964]: DEBUG nova.compute.provider_tree [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.876323] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 962.878104] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041375, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.891032] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Successfully created port: da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.920069] env[61964]: DEBUG nova.network.neutron [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.990260] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-f983316e-e012-41fd-bf20-7d9cc5accd71" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.991065] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance network_info: |[{"id": "a70cd208-c6e3-44de-9b66-dd509b9b8a1b", "address": "fa:16:3e:9a:d4:3a", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa70cd208-c6", "ovs_interfaceid": "a70cd208-c6e3-44de-9b66-dd509b9b8a1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 962.993020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:d4:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a70cd208-c6e3-44de-9b66-dd509b9b8a1b', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.000232] env[61964]: DEBUG oslo.service.loopingcall [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.001561] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.004017] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 963.004017] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-497ab183-8aa8-4ae0-91a5-8506d7ee737a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.029061] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.029061] env[61964]: value = "task-1041376" [ 963.029061] env[61964]: _type = "Task" [ 963.029061] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.041136] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041376, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.051883] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d1541cb-746a-4512-a7fc-7c4bd224b1f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.062044] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53b152d-81b8-40ec-91cb-811bb96d0d4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.094356] env[61964]: DEBUG nova.compute.manager [req-edc7268f-059d-4014-a9a6-415528480e2d req-d81eb00c-bbce-4074-9280-b4e9d6af51cf service nova] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Detach interface failed, port_id=edda620f-1259-4fbb-afa6-48aef4eda40b, reason: Instance 63793ce6-d511-403e-8a4b-cad8c4157449 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 963.243690] env[61964]: DEBUG nova.scheduler.client.report [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.376556] env[61964]: DEBUG oslo_vmware.api [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041375, 'name': PowerOnVM_Task, 'duration_secs': 0.803506} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.376889] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 963.377119] env[61964]: INFO nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Took 8.04 seconds to spawn the instance on the hypervisor. [ 963.377313] env[61964]: DEBUG nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 963.378283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f13f9cf-aeb5-4479-9ac7-da1dcd038c00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.392113] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 963.392113] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230545', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'name': 'volume-9930fa82-359c-40a3-9658-852e13df340f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a14232f-5cc6-4856-9291-eb0a1bd2ab69', 'attached_at': '', 'detached_at': '', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'serial': '9930fa82-359c-40a3-9658-852e13df340f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 963.397524] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e10cd6-c2ab-4082-8c30-620bf01d8bfc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.417932] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b007fb-8182-462b-aa68-5a5d08dd5423 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.424273] env[61964]: INFO nova.compute.manager [-] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Took 1.65 seconds to deallocate network for instance. [ 963.452402] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] volume-9930fa82-359c-40a3-9658-852e13df340f/volume-9930fa82-359c-40a3-9658-852e13df340f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.457091] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f6aba39-d5bf-4bbf-b817-620438f32bbe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.482936] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 963.482936] env[61964]: value = "task-1041377" [ 963.482936] env[61964]: _type = "Task" [ 963.482936] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.494896] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041377, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.538854] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041376, 'name': CreateVM_Task, 'duration_secs': 0.380337} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.539230] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.540095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.540400] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.541023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.541218] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d51ee1d-fda9-4ee6-b4c5-823906d9520c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.549122] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 963.549122] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527db035-06f0-ea76-7e16-84c8c06f1988" [ 963.549122] env[61964]: _type = "Task" [ 963.549122] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.556377] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527db035-06f0-ea76-7e16-84c8c06f1988, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.750560] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.756379] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.309s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.756379] env[61964]: DEBUG nova.objects.instance [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'resources' on Instance uuid d8f24177-b695-4d40-958d-9272b4f683cc {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.782349] env[61964]: INFO nova.scheduler.client.report [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Deleted allocations for instance f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8 [ 963.905214] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 963.914261] env[61964]: INFO nova.compute.manager [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Took 29.37 seconds to build instance. [ 963.937870] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 963.938236] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.938487] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 963.938747] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.938963] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 963.939191] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 963.939460] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 963.939683] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 963.939914] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 963.940149] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 963.940386] env[61964]: DEBUG nova.virt.hardware [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 963.941527] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e520af9-41e8-44f8-9eaa-712c374d784a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.950142] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a349d2df-112a-428a-989f-4cd10ebdb06f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.978245] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.991705] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041377, 'name': ReconfigVM_Task, 'duration_secs': 0.405066} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.992075] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfigured VM instance instance-00000053 to attach disk [datastore1] volume-9930fa82-359c-40a3-9658-852e13df340f/volume-9930fa82-359c-40a3-9658-852e13df340f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.997233] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2eae6d07-db67-4eef-8397-3a53c2993aeb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.012159] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 964.012159] env[61964]: value = "task-1041378" [ 964.012159] env[61964]: _type = "Task" [ 964.012159] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.021161] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.057945] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527db035-06f0-ea76-7e16-84c8c06f1988, 'name': SearchDatastore_Task, 'duration_secs': 0.012148} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.058179] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.058552] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.058673] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.058826] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.059020] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.059297] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2571d6c5-d778-497f-8678-1ea7d23dbad2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.068469] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.068763] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 964.069584] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba395495-8047-4cb2-9b3f-f8147288ab9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.075589] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 964.075589] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb9180-8ce0-78eb-45f2-ceb301892e1a" [ 964.075589] env[61964]: _type = "Task" [ 964.075589] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.086207] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb9180-8ce0-78eb-45f2-ceb301892e1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.298538] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8e9cff8c-9d5b-4360-a268-fabbe9bd9b6f tempest-ServersTestManualDisk-968274791 tempest-ServersTestManualDisk-968274791-project-member] Lock "f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.566s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.416266] env[61964]: DEBUG oslo_concurrency.lockutils [None req-16893b9e-9d50-4925-a71b-13329fdce472 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.880s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.476050] env[61964]: DEBUG nova.compute.manager [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 964.476050] env[61964]: DEBUG nova.compute.manager [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing instance network info cache due to event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 964.476050] env[61964]: DEBUG oslo_concurrency.lockutils [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.476050] env[61964]: DEBUG oslo_concurrency.lockutils [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.476050] env[61964]: DEBUG nova.network.neutron [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.527414] env[61964]: DEBUG nova.compute.manager [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Received event network-vif-plugged-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 964.527685] env[61964]: DEBUG oslo_concurrency.lockutils [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.528049] env[61964]: DEBUG oslo_concurrency.lockutils [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.528049] env[61964]: DEBUG oslo_concurrency.lockutils [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.529183] env[61964]: DEBUG nova.compute.manager [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] No waiting events found dispatching network-vif-plugged-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 964.529608] env[61964]: WARNING nova.compute.manager [req-75caa35e-0492-4c0b-b0be-4bd0bac2a046 req-e8ebbd5a-f654-4b87-8475-59f72c9ea6a8 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Received unexpected event network-vif-plugged-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 for instance with vm_state building and task_state spawning. [ 964.539527] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.539974] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d654f56f-eed2-401b-97f9-830e8845a45d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.548873] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b019bc-944c-4eb4-a4a0-3d7011283bb2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.583071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac43c0c-71a3-4e61-a188-4b6d6f76baec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.591434] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb9180-8ce0-78eb-45f2-ceb301892e1a, 'name': SearchDatastore_Task, 'duration_secs': 0.009411} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.594763] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7726d4ca-c543-43ae-a611-3b451a0aad9d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.598681] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9e737c-d4a4-4b91-95e7-adb8af949011 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.605259] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 964.605259] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521f0f0a-04b2-669e-35ad-34fe0e1bcdc9" [ 964.605259] env[61964]: _type = "Task" [ 964.605259] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.613937] env[61964]: DEBUG nova.compute.provider_tree [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.623570] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521f0f0a-04b2-669e-35ad-34fe0e1bcdc9, 'name': SearchDatastore_Task, 'duration_secs': 0.011302} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.624666] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.624768] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] f983316e-e012-41fd-bf20-7d9cc5accd71/f983316e-e012-41fd-bf20-7d9cc5accd71.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.625027] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f33f0372-8839-4888-ab87-a6166957c31f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.637911] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 964.637911] env[61964]: value = "task-1041379" [ 964.637911] env[61964]: _type = "Task" [ 964.637911] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.645618] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.747012] env[61964]: DEBUG nova.compute.manager [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 964.747277] env[61964]: DEBUG nova.compute.manager [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 964.747569] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.747743] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.748011] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 965.033274] env[61964]: DEBUG oslo_vmware.api [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041378, 'name': ReconfigVM_Task, 'duration_secs': 0.996818} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.033274] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230545', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'name': 'volume-9930fa82-359c-40a3-9658-852e13df340f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a14232f-5cc6-4856-9291-eb0a1bd2ab69', 'attached_at': '', 'detached_at': '', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'serial': '9930fa82-359c-40a3-9658-852e13df340f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 965.104385] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Successfully updated port: da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.119708] env[61964]: DEBUG nova.scheduler.client.report [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.141017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.141017] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.141017] env[61964]: DEBUG nova.objects.instance [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'flavor' on Instance uuid 47378856-4cbf-4cf6-aecd-ae935885df3d {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.157339] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476436} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.157339] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] f983316e-e012-41fd-bf20-7d9cc5accd71/f983316e-e012-41fd-bf20-7d9cc5accd71.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.157339] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.157339] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-daab5102-e670-4713-9f5a-6434a9350364 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.163841] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 965.163841] env[61964]: value = "task-1041380" [ 965.163841] env[61964]: _type = "Task" [ 965.163841] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.176422] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041380, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.228583] env[61964]: DEBUG nova.network.neutron [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updated VIF entry in instance network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.228974] env[61964]: DEBUG nova.network.neutron [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.487019] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.487019] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.606173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.606331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.606485] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.625352] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.873s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.627748] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.912s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.628092] env[61964]: DEBUG nova.objects.instance [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid 2c77a07d-c949-4aef-8d6f-132969d58cee {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.650278] env[61964]: INFO nova.scheduler.client.report [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted allocations for instance d8f24177-b695-4d40-958d-9272b4f683cc [ 965.674386] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041380, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177975} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.676892] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.677859] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00dc01cf-65fd-438a-8be1-dd1bfe1514c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.702138] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] f983316e-e012-41fd-bf20-7d9cc5accd71/f983316e-e012-41fd-bf20-7d9cc5accd71.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.702587] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87e935a6-de6b-48c0-b45c-d9a148353286 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.727446] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 965.727446] env[61964]: value = "task-1041381" [ 965.727446] env[61964]: _type = "Task" [ 965.727446] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.731480] env[61964]: DEBUG oslo_concurrency.lockutils [req-049ea1ac-7c20-444e-b2a5-42f3ce3114a8 req-51ea62ad-7f38-44d4-a147-a264c8e18113 service nova] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.737869] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041381, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.923974] env[61964]: DEBUG nova.objects.instance [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'pci_requests' on Instance uuid 47378856-4cbf-4cf6-aecd-ae935885df3d {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.923974] env[61964]: DEBUG nova.compute.manager [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Stashing vm_state: active {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 965.990305] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.990689] env[61964]: DEBUG nova.compute.manager [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 965.990930] env[61964]: DEBUG nova.compute.manager [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing instance network info cache due to event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 965.991246] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.991445] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.991667] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.079988] env[61964]: DEBUG nova.objects.instance [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.163080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d8ece4f5-613d-4068-b822-9368609a5d11 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "d8f24177-b695-4d40-958d-9272b4f683cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.133s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.194279] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 966.243913] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041381, 'name': ReconfigVM_Task, 'duration_secs': 0.299877} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.244317] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Reconfigured VM instance instance-0000005c to attach disk [datastore1] f983316e-e012-41fd-bf20-7d9cc5accd71/f983316e-e012-41fd-bf20-7d9cc5accd71.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.244981] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd281098-bc72-417d-b0f5-6cb8f47db977 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.252567] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 966.252567] env[61964]: value = "task-1041382" [ 966.252567] env[61964]: _type = "Task" [ 966.252567] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.263364] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041382, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.345937] env[61964]: DEBUG nova.objects.base [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Object Instance<47378856-4cbf-4cf6-aecd-ae935885df3d> lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 966.346828] env[61964]: DEBUG nova.network.neutron [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 966.427822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.450301] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d6d18a-e056-49e1-8c01-de39439b6a97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.462207] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7506eaf9-c369-4c11-afe8-c3e4921dac9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.508590] env[61964]: DEBUG nova.policy [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.510910] env[61964]: DEBUG nova.network.neutron [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.512503] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d28b75-b8fc-48a0-8491-03c62f3679b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.521165] env[61964]: DEBUG nova.compute.manager [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 966.521165] env[61964]: DEBUG nova.compute.manager [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 966.521165] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.521165] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.521165] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.529120] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6fa9fa-e9e9-4810-b297-b1a053b584dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.544029] env[61964]: DEBUG nova.compute.provider_tree [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.586164] env[61964]: DEBUG oslo_concurrency.lockutils [None req-49ec3017-9f69-4255-982d-b96027d50b77 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.826s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.766014] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041382, 'name': Rename_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.002136] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updated VIF entry in instance network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.002341] env[61964]: DEBUG nova.network.neutron [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.020146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.020479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.020739] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.020955] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.021171] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.022937] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.024389] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Instance network_info: |[{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 967.024389] env[61964]: INFO nova.compute.manager [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Terminating instance [ 967.029792] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:50:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da9bd08d-0a86-4d20-82b3-0160a3cf26b2', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.038383] env[61964]: DEBUG oslo.service.loopingcall [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.040799] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 967.044506] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36c7b55e-5f43-409d-9fe4-d80a1037ba47 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.064574] env[61964]: DEBUG nova.scheduler.client.report [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.076825] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.076825] env[61964]: value = "task-1041383" [ 967.076825] env[61964]: _type = "Task" [ 967.076825] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.086198] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041383, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.263581] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041382, 'name': Rename_Task, 'duration_secs': 0.872136} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.263907] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 967.264198] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25053e2a-d1e4-48b8-82c4-f7e04b953b53 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.271128] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 967.271128] env[61964]: value = "task-1041384" [ 967.271128] env[61964]: _type = "Task" [ 967.271128] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.279520] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.410201] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.410643] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.505534] env[61964]: DEBUG oslo_concurrency.lockutils [req-e82a7832-e3f7-4146-8ade-b6fa1e269258 req-706ea578-383b-4693-b771-b27d9b4559c0 service nova] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.541706] env[61964]: DEBUG nova.compute.manager [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 967.541987] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 967.546018] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e06256f-66b9-414e-b9c2-ab30fe11aa01 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.551829] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 967.551829] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69f355a4-1d39-4619-88b3-4bd8cdaf9ccf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.558561] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 967.558561] env[61964]: value = "task-1041385" [ 967.558561] env[61964]: _type = "Task" [ 967.558561] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.567784] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.569895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.573399] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.572s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.573724] env[61964]: DEBUG nova.objects.instance [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lazy-loading 'resources' on Instance uuid 58f1ad74-ea72-45c8-b98c-c85674989342 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.590846] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041383, 'name': CreateVM_Task, 'duration_secs': 0.456533} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.591566] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 967.592309] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.592625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.592891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.593570] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c28b2db1-da0d-4c0d-9a6e-4bbd48ebd136 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.597517] env[61964]: INFO nova.scheduler.client.report [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance 2c77a07d-c949-4aef-8d6f-132969d58cee [ 967.599906] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 967.599906] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525e2388-08b2-d6ce-a2ad-a9ce5fcf7f45" [ 967.599906] env[61964]: _type = "Task" [ 967.599906] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.616676] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]525e2388-08b2-d6ce-a2ad-a9ce5fcf7f45, 'name': SearchDatastore_Task, 'duration_secs': 0.010724} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.616974] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.617235] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.617531] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.617652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.617803] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.618408] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b46836ca-bddc-40bb-b6ea-8f80b78954d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.628163] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.628163] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.628273] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dbe6294-7fdd-44fe-a6c4-7a1bec0e89da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.633986] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 967.633986] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c58ae7-d81e-aaeb-582a-fe87ceeda814" [ 967.633986] env[61964]: _type = "Task" [ 967.633986] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.642307] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c58ae7-d81e-aaeb-582a-fe87ceeda814, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.780987] env[61964]: DEBUG oslo_vmware.api [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041384, 'name': PowerOnVM_Task, 'duration_secs': 0.502408} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.781302] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.781515] env[61964]: INFO nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Took 8.15 seconds to spawn the instance on the hypervisor. [ 967.781708] env[61964]: DEBUG nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 967.782502] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1030a4fd-3643-41db-a5a2-26b19dd84cab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.915086] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.915086] env[61964]: DEBUG nova.compute.manager [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Received event network-changed-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 967.915086] env[61964]: DEBUG nova.compute.manager [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Refreshing instance network info cache due to event network-changed-da9bd08d-0a86-4d20-82b3-0160a3cf26b2. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 967.915086] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.915086] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.915364] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Refreshing network info cache for port da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.019741] env[61964]: DEBUG nova.network.neutron [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Successfully updated port: e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.069608] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041385, 'name': PowerOffVM_Task, 'duration_secs': 0.256869} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.069883] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.070070] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.070328] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66060222-a98e-41d2-b234-75345aee98c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.112859] env[61964]: DEBUG oslo_concurrency.lockutils [None req-48b163ae-a627-4c79-a141-4ccd56929473 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "2c77a07d-c949-4aef-8d6f-132969d58cee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.759s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.138748] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.138992] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.139198] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleting the datastore file [datastore1] 003a5b74-2b8d-4e14-a7ee-db8006f81dfa {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.140953] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba38f595-b333-4130-85ea-8004a362f65a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.152579] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "90679b1f-2c9c-4049-bfb9-db573220163b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.152850] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.154327] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c58ae7-d81e-aaeb-582a-fe87ceeda814, 'name': SearchDatastore_Task, 'duration_secs': 0.008785} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.156565] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6c46ac4-60d8-47d6-bbfd-2eae22233d2f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.159627] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 968.159627] env[61964]: value = "task-1041387" [ 968.159627] env[61964]: _type = "Task" [ 968.159627] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.165842] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 968.165842] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521b00fa-01a3-8bd2-0941-d4b2feef60ae" [ 968.165842] env[61964]: _type = "Task" [ 968.165842] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.176612] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041387, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.181918] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521b00fa-01a3-8bd2-0941-d4b2feef60ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.291320] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bdf17a-572a-4055-a6c5-471b384bdf8a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.304167] env[61964]: INFO nova.compute.manager [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Took 15.68 seconds to build instance. [ 968.305552] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd5f5ce-cdc2-4010-8b1e-c136f0798853 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.338224] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddca8556-1161-40c1-b054-237f3aea2c8c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.345995] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78503f52-41c7-4c39-8612-ac7d34363e31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.359733] env[61964]: DEBUG nova.compute.provider_tree [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.526072] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.526325] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.526601] env[61964]: DEBUG nova.network.neutron [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 968.546741] env[61964]: DEBUG nova.compute.manager [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 968.547039] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.547211] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.547427] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.547551] env[61964]: DEBUG nova.compute.manager [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] No waiting events found dispatching network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 968.547721] env[61964]: WARNING nova.compute.manager [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received unexpected event network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 for instance with vm_state active and task_state None. [ 968.547884] env[61964]: DEBUG nova.compute.manager [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 968.548055] env[61964]: DEBUG nova.compute.manager [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-e678489f-0505-421c-893c-43523e7e29f7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 968.548282] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.586166] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "4d9c24cf-e11b-4d95-ad41-538d91279689" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.586483] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.586706] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.586910] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.587159] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.589389] env[61964]: INFO nova.compute.manager [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Terminating instance [ 968.617558] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updated VIF entry in instance network info cache for port da9bd08d-0a86-4d20-82b3-0160a3cf26b2. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 968.617936] env[61964]: DEBUG nova.network.neutron [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.655335] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 968.669869] env[61964]: DEBUG oslo_vmware.api [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041387, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360207} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.674054] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.674231] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.674325] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.674454] env[61964]: INFO nova.compute.manager [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 968.674697] env[61964]: DEBUG oslo.service.loopingcall [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.674901] env[61964]: DEBUG nova.compute.manager [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 968.674996] env[61964]: DEBUG nova.network.neutron [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 968.682083] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521b00fa-01a3-8bd2-0941-d4b2feef60ae, 'name': SearchDatastore_Task, 'duration_secs': 0.018446} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.682355] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.682602] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.682872] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb7c53f7-97ab-4032-98b1-48194dd8a106 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.688936] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 968.688936] env[61964]: value = "task-1041388" [ 968.688936] env[61964]: _type = "Task" [ 968.688936] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.697855] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041388, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.810505] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aee91ba2-fe06-491a-9b8b-b1ce3e05b9a8 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.195s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.863590] env[61964]: DEBUG nova.scheduler.client.report [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 969.065474] env[61964]: WARNING nova.network.neutron [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] 48dd64e7-3659-4104-8273-6f58f8d3eb44 already exists in list: networks containing: ['48dd64e7-3659-4104-8273-6f58f8d3eb44']. ignoring it [ 969.093214] env[61964]: DEBUG nova.compute.manager [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 969.093476] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 969.094549] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6b9544-9ee6-4980-8717-1f1bfc59cff4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.104518] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.104878] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13a4838a-9e31-45d0-a889-a207c5212b9f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.116344] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 969.116344] env[61964]: value = "task-1041389" [ 969.116344] env[61964]: _type = "Task" [ 969.116344] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.120956] env[61964]: DEBUG oslo_concurrency.lockutils [req-996fe0e0-72cc-472d-babb-47eea3c75860 req-5539093e-d26e-46a1-8425-eaa7cf9f8014 service nova] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.127067] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.187822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.200533] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041388, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.369610] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.372034] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.394s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.372034] env[61964]: DEBUG nova.objects.instance [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lazy-loading 'resources' on Instance uuid 63793ce6-d511-403e-8a4b-cad8c4157449 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.391815] env[61964]: INFO nova.scheduler.client.report [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Deleted allocations for instance 58f1ad74-ea72-45c8-b98c-c85674989342 [ 969.400793] env[61964]: DEBUG nova.network.neutron [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e678489f-0505-421c-893c-43523e7e29f7", "address": "fa:16:3e:3a:ec:64", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape678489f-05", "ovs_interfaceid": "e678489f-0505-421c-893c-43523e7e29f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.426986] env[61964]: DEBUG nova.network.neutron [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.626995] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041389, 'name': PowerOffVM_Task, 'duration_secs': 0.215758} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.627341] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.627504] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 969.627752] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2011a129-0ade-4ba6-b9fc-f91ffa6f1865 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.700107] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041388, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596276} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.700389] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 969.700608] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.700854] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d217fac-e8a9-436c-98ad-5c0fe0fbcd04 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.706572] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 969.706572] env[61964]: value = "task-1041391" [ 969.706572] env[61964]: _type = "Task" [ 969.706572] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.713372] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.861495] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 969.861715] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 969.861908] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore2] 4d9c24cf-e11b-4d95-ad41-538d91279689 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.863944] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37c198f6-deef-4c65-a3c2-21282ec6443c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.868645] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 969.868645] env[61964]: value = "task-1041392" [ 969.868645] env[61964]: _type = "Task" [ 969.868645] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.878933] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.900784] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d866c03a-61e6-4fe7-a2bc-47ce5907fba8 tempest-ImagesTestJSON-116688785 tempest-ImagesTestJSON-116688785-project-member] Lock "58f1ad74-ea72-45c8-b98c-c85674989342" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.580s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.903084] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.904105] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.904105] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.906632] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.906829] env[61964]: DEBUG nova.network.neutron [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.908376] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de486813-3739-4a0a-9641-a10685f1e9dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.925246] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 969.925457] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.925620] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 969.925836] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.926031] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 969.926230] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 969.926444] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 969.926608] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 969.926780] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 969.926940] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 969.927135] env[61964]: DEBUG nova.virt.hardware [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 969.933793] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfiguring VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 969.937670] env[61964]: INFO nova.compute.manager [-] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Took 1.26 seconds to deallocate network for instance. [ 969.937915] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e63c09f-c643-43e6-b824-5697ada4b919 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.957594] env[61964]: DEBUG oslo_vmware.api [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 969.957594] env[61964]: value = "task-1041393" [ 969.957594] env[61964]: _type = "Task" [ 969.957594] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.969025] env[61964]: DEBUG oslo_vmware.api [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041393, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.094312] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a841858-e1c6-4f60-b032-36aaaaf44c34 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.097354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f983316e-e012-41fd-bf20-7d9cc5accd71" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.097592] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.097775] env[61964]: DEBUG nova.compute.manager [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 970.098657] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2b97a0-6199-472c-bd03-df33b90815e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.105401] env[61964]: DEBUG nova.compute.manager [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 970.106026] env[61964]: DEBUG nova.objects.instance [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'flavor' on Instance uuid f983316e-e012-41fd-bf20-7d9cc5accd71 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.110606] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef962757-bd19-4c1e-9e57-04749b6c415d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.143316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d44bac1-09e0-45ef-b493-109f8a2a4672 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.153149] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7763f779-dd4a-4044-981b-76b4616974c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.169905] env[61964]: DEBUG nova.compute.provider_tree [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.217257] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.382032] env[61964]: DEBUG oslo_vmware.api [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.485175} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.382032] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.382032] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 970.382032] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 970.382032] env[61964]: INFO nova.compute.manager [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Took 1.29 seconds to destroy the instance on the hypervisor. [ 970.382032] env[61964]: DEBUG oslo.service.loopingcall [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.382032] env[61964]: DEBUG nova.compute.manager [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 970.382032] env[61964]: DEBUG nova.network.neutron [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.457236] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.469729] env[61964]: DEBUG oslo_vmware.api [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.676026] env[61964]: DEBUG nova.scheduler.client.report [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.717526] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.666386} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.718040] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.719026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad9fead-8a71-415a-8645-5394e93107d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.723181] env[61964]: DEBUG nova.network.neutron [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port e678489f-0505-421c-893c-43523e7e29f7. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 970.723761] env[61964]: DEBUG nova.network.neutron [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e678489f-0505-421c-893c-43523e7e29f7", "address": "fa:16:3e:3a:ec:64", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape678489f-05", "ovs_interfaceid": "e678489f-0505-421c-893c-43523e7e29f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.751636] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.752913] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bede0351-cdec-4562-9ab0-c96f6a48d066 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.775883] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 970.775883] env[61964]: value = "task-1041394" [ 970.775883] env[61964]: _type = "Task" [ 970.775883] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.785334] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041394, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.855174] env[61964]: DEBUG nova.compute.manager [req-34806fa5-8a3b-43d6-a709-c91d4e6fa5f6 req-3c271651-a4c4-47c3-b6a4-6596eaf5e91d service nova] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Received event network-vif-deleted-6b326930-a132-4e0d-831c-c44b7ac08c9f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 970.910325] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.910640] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.910867] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.911089] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.911280] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.914184] env[61964]: INFO nova.compute.manager [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Terminating instance [ 970.971255] env[61964]: DEBUG oslo_vmware.api [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.114716] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.115468] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99bf085b-398b-4481-b11c-9d2978312843 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.123363] env[61964]: DEBUG oslo_vmware.api [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 971.123363] env[61964]: value = "task-1041395" [ 971.123363] env[61964]: _type = "Task" [ 971.123363] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.134555] env[61964]: DEBUG oslo_vmware.api [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.180098] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.183829] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.756s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.215727] env[61964]: INFO nova.scheduler.client.report [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Deleted allocations for instance 63793ce6-d511-403e-8a4b-cad8c4157449 [ 971.227582] env[61964]: DEBUG oslo_concurrency.lockutils [req-92137464-da0e-43be-b26b-bcb8a891e628 req-b0c91141-a05a-4182-a49f-7e1d59d587b8 service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.272138] env[61964]: DEBUG nova.network.neutron [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.286174] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041394, 'name': ReconfigVM_Task, 'duration_secs': 0.506632} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.287035] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.287600] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e2e6d66-7110-494b-ba23-958999741529 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.294862] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 971.294862] env[61964]: value = "task-1041396" [ 971.294862] env[61964]: _type = "Task" [ 971.294862] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.303823] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041396, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.423283] env[61964]: DEBUG nova.compute.manager [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 971.423283] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.423843] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6a6838-ed32-4d68-a829-01c85b41ac29 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.433601] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.433857] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c15792a7-aa23-4e71-806e-37b59a50ae83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.439842] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 971.439842] env[61964]: value = "task-1041397" [ 971.439842] env[61964]: _type = "Task" [ 971.439842] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.447435] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.467987] env[61964]: DEBUG oslo_vmware.api [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041393, 'name': ReconfigVM_Task, 'duration_secs': 1.17867} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.468568] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.468792] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfigured VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 971.633898] env[61964]: DEBUG oslo_vmware.api [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041395, 'name': PowerOffVM_Task, 'duration_secs': 0.442254} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.634216] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.634421] env[61964]: DEBUG nova.compute.manager [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 971.635234] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01a087e-2fe8-47ea-8f71-20fdee9e254a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.692207] env[61964]: INFO nova.compute.claims [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.726397] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1e22b71e-c551-4b96-a15e-8d9c2c92531a tempest-ServersNegativeTestJSON-1289611832 tempest-ServersNegativeTestJSON-1289611832-project-member] Lock "63793ce6-d511-403e-8a4b-cad8c4157449" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.592s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.774519] env[61964]: INFO nova.compute.manager [-] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Took 1.39 seconds to deallocate network for instance. [ 971.804812] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041396, 'name': Rename_Task, 'duration_secs': 0.247978} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.805117] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.805372] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f85b5571-d43c-413e-b1e0-60b046058760 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.811628] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 971.811628] env[61964]: value = "task-1041398" [ 971.811628] env[61964]: _type = "Task" [ 971.811628] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.819668] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041398, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.951110] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041397, 'name': PowerOffVM_Task, 'duration_secs': 0.178328} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.951513] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.951735] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.952015] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0f6561c-79e6-4c8c-b09a-3acc600ace31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.973305] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df621d37-5781-42cc-8379-af13b3dee137 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.835s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.031371] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.031642] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.031898] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleting the datastore file [datastore1] 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.032209] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-326758e3-3d18-4fa7-a06e-8ba50085892e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.040141] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 972.040141] env[61964]: value = "task-1041400" [ 972.040141] env[61964]: _type = "Task" [ 972.040141] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.048198] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.147104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b2d314c2-bffe-4408-a565-83f1dd377792 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.199586] env[61964]: INFO nova.compute.resource_tracker [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating resource usage from migration 9a2fc428-7e53-4418-89a7-f123879cb767 [ 972.282809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.325898] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041398, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.406825] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38083dd-9e0b-4205-9317-e0ff835ee018 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.417712] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef753d1-bf0d-4a40-ae52-cfba17e872ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.447223] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dcd096-b8d6-4c4e-a6fb-e3a9373d5ba0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.455389] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612c501b-3522-48ec-bddd-f5ad70c176bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.469403] env[61964]: DEBUG nova.compute.provider_tree [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.552391] env[61964]: DEBUG oslo_vmware.api [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326937} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.552647] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.552795] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.552974] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.553181] env[61964]: INFO nova.compute.manager [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 972.553488] env[61964]: DEBUG oslo.service.loopingcall [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.553687] env[61964]: DEBUG nova.compute.manager [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 972.553781] env[61964]: DEBUG nova.network.neutron [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.825516] env[61964]: DEBUG oslo_vmware.api [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041398, 'name': PowerOnVM_Task, 'duration_secs': 0.768199} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.825840] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 972.828747] env[61964]: INFO nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Took 8.92 seconds to spawn the instance on the hypervisor. [ 972.828747] env[61964]: DEBUG nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 972.828747] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95f8008-7998-413a-a12a-5030801ebbc5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.973133] env[61964]: DEBUG nova.scheduler.client.report [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.005643] env[61964]: DEBUG nova.compute.manager [req-38e9ed4f-d8dd-4f36-9fc2-b28b8c463fa7 req-8c8c935b-dc6c-4dff-9587-5bcde6765517 service nova] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Received event network-vif-deleted-36f6ac10-084a-4796-9f7d-cc945a356b01 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 973.348716] env[61964]: INFO nova.compute.manager [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Took 19.97 seconds to build instance. [ 973.390907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f983316e-e012-41fd-bf20-7d9cc5accd71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.390907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.390907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.390907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.390907] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.391661] env[61964]: INFO nova.compute.manager [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Terminating instance [ 973.478657] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.295s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.479556] env[61964]: INFO nova.compute.manager [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Migrating [ 973.489347] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.299s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.489347] env[61964]: INFO nova.compute.claims [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 973.572040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.572040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.850841] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4a4f412-8367-4f13-9e08-694adf2614df tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.480s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.895830] env[61964]: DEBUG nova.compute.manager [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 973.896117] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.897471] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f2b198-f6c7-4f02-bceb-567cc32bcb5b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.910197] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 973.910479] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff1450ac-1ac5-471a-a6a1-4bc9ff332b1b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.982415] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 973.982643] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 973.982833] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore1] f983316e-e012-41fd-bf20-7d9cc5accd71 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.983126] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdbc9267-f1e8-46a8-9591-1ca2959a412d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.989875] env[61964]: DEBUG oslo_vmware.api [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 973.989875] env[61964]: value = "task-1041402" [ 973.989875] env[61964]: _type = "Task" [ 973.989875] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.997603] env[61964]: DEBUG oslo_vmware.api [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.005450] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.006084] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.006084] env[61964]: DEBUG nova.network.neutron [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 974.056881] env[61964]: DEBUG nova.compute.manager [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Received event network-changed-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 974.057140] env[61964]: DEBUG nova.compute.manager [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Refreshing instance network info cache due to event network-changed-da9bd08d-0a86-4d20-82b3-0160a3cf26b2. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 974.058909] env[61964]: DEBUG oslo_concurrency.lockutils [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.058909] env[61964]: DEBUG oslo_concurrency.lockutils [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.058909] env[61964]: DEBUG nova.network.neutron [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Refreshing network info cache for port da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.074680] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.074680] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.075630] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc62ea72-8f71-42a7-8824-8708b05c8eb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.097718] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755c5c6b-2dd1-4324-bc09-4045b105a1e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.131350] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfiguring VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 974.131974] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecf2998c-2a93-4715-aa4b-26055445b0fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.152355] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 974.152355] env[61964]: value = "task-1041403" [ 974.152355] env[61964]: _type = "Task" [ 974.152355] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.164385] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.176502] env[61964]: DEBUG nova.network.neutron [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.498902] env[61964]: DEBUG oslo_vmware.api [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157842} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.501553] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.501794] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 974.502023] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 974.502222] env[61964]: INFO nova.compute.manager [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Took 0.61 seconds to destroy the instance on the hypervisor. [ 974.502476] env[61964]: DEBUG oslo.service.loopingcall [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.502722] env[61964]: DEBUG nova.compute.manager [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 974.502828] env[61964]: DEBUG nova.network.neutron [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 974.662416] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.680788] env[61964]: INFO nova.compute.manager [-] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Took 2.13 seconds to deallocate network for instance. [ 974.745966] env[61964]: DEBUG nova.network.neutron [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.763131] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809f7a2c-4af9-46e0-af41-dee251485625 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.771296] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a21c090-14c1-43da-a70c-1b57ca0fffb5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.805108] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80639da0-affc-47ed-ab03-71989e6058d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.817811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2733a7-4650-4534-a0aa-41d5695ddcb9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.833333] env[61964]: DEBUG nova.compute.provider_tree [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.101474] env[61964]: DEBUG nova.compute.manager [req-0cf98e44-335b-4a56-b6cf-772f55dd9868 req-c923c9ce-484a-4048-82b8-fc739d2d7063 service nova] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Received event network-vif-deleted-3a986bcd-6584-4440-bf01-7db0dff75b2f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 975.169767] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.191030] env[61964]: DEBUG nova.network.neutron [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updated VIF entry in instance network info cache for port da9bd08d-0a86-4d20-82b3-0160a3cf26b2. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 975.191439] env[61964]: DEBUG nova.network.neutron [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.194871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.248390] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.339283] env[61964]: DEBUG nova.scheduler.client.report [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.647696] env[61964]: DEBUG nova.network.neutron [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.665352] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.697879] env[61964]: DEBUG oslo_concurrency.lockutils [req-5051644d-6390-43ce-88a7-307174e8288b req-8152d03f-8b20-4041-9007-8e440db17e03 service nova] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.843103] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.844937] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 975.846458] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.389s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.847317] env[61964]: DEBUG nova.objects.instance [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'resources' on Instance uuid 003a5b74-2b8d-4e14-a7ee-db8006f81dfa {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.151523] env[61964]: INFO nova.compute.manager [-] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Took 1.65 seconds to deallocate network for instance. [ 976.169069] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.350160] env[61964]: DEBUG nova.compute.utils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 976.356790] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 976.356790] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 976.421023] env[61964]: DEBUG nova.policy [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf85b27f46c94956982ca3c713aef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b79a033963e04e02a98df18f8429263f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 976.595371] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9af46ce-a0ec-4bf0-8bcf-8c2a24a9d757 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.606547] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80029e65-3e72-466b-88c4-669781475e78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.650299] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03d21a7-20b7-46e1-be08-88c6c0e7e6fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.663127] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f032b3-3b4c-44a7-a60a-c2211e24fe25 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.668235] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.674559] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.683594] env[61964]: DEBUG nova.compute.provider_tree [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.738786] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Successfully created port: f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.749587] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.749895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.767235] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a7f452-b734-420b-b0ac-d9574131c9ad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.789264] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 976.857916] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 977.131410] env[61964]: DEBUG nova.compute.manager [req-fb9fcb93-3a3b-4925-bd6d-edc7ebfe8059 req-2299d811-3000-4951-87b5-b31b849a25cd service nova] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Received event network-vif-deleted-a70cd208-c6e3-44de-9b66-dd509b9b8a1b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 977.169408] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.190465] env[61964]: DEBUG nova.scheduler.client.report [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.252900] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 977.296459] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.297082] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e90c5a2-aeb3-4e6a-aded-ae7850a2abb2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.305200] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 977.305200] env[61964]: value = "task-1041404" [ 977.305200] env[61964]: _type = "Task" [ 977.305200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.314457] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.669280] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.698250] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.700668] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.418s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.700929] env[61964]: DEBUG nova.objects.instance [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid 4d9c24cf-e11b-4d95-ad41-538d91279689 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.722482] env[61964]: INFO nova.scheduler.client.report [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted allocations for instance 003a5b74-2b8d-4e14-a7ee-db8006f81dfa [ 977.773913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.815694] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041404, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.868578] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 977.895290] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 977.895559] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.895726] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 977.895917] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.896082] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 977.896240] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 977.896470] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 977.896654] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 977.896833] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 977.897413] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 977.897413] env[61964]: DEBUG nova.virt.hardware [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 977.898124] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42544c9b-d807-4d78-ac57-cc973c68158f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.906323] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfec733d-a8c1-44ce-b748-0f1f4ca6644e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.171772] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.230012] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a4e54b54-29b6-4677-a88f-326b6de180cf tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "003a5b74-2b8d-4e14-a7ee-db8006f81dfa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.209s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.294853] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.295332] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.317710] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041404, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.445167] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Successfully updated port: f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 978.473799] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b808ec77-f42e-488f-a078-453862332952 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.481913] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f9ef55-edab-4a40-ae1f-bb902349178e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.513926] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b1edb0-0c32-47b4-80dc-7a2b151c796a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.521499] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a36f489-9820-46bf-b142-0fe67280345f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.537153] env[61964]: DEBUG nova.compute.provider_tree [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.669154] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.805315] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.805518] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 978.815839] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041404, 'name': PowerOffVM_Task, 'duration_secs': 1.253786} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.816110] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.816301] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 978.893015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "f5acabe4-674b-4c4c-85b4-b9270926017b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.893317] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.893536] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.893731] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.893906] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.896121] env[61964]: INFO nova.compute.manager [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Terminating instance [ 978.948825] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.948959] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.949118] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.041358] env[61964]: DEBUG nova.scheduler.client.report [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.157971] env[61964]: DEBUG nova.compute.manager [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Received event network-vif-plugged-f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 979.158207] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Acquiring lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.158464] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.158609] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.159041] env[61964]: DEBUG nova.compute.manager [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] No waiting events found dispatching network-vif-plugged-f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 979.159284] env[61964]: WARNING nova.compute.manager [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Received unexpected event network-vif-plugged-f0e7e755-ae40-4d0d-bf57-2706d53067d6 for instance with vm_state building and task_state spawning. [ 979.159826] env[61964]: DEBUG nova.compute.manager [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Received event network-changed-f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 979.159826] env[61964]: DEBUG nova.compute.manager [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Refreshing instance network info cache due to event network-changed-f0e7e755-ae40-4d0d-bf57-2706d53067d6. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 979.159826] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Acquiring lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.170924] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.322807] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 979.324108] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 979.324778] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 979.324778] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 979.324778] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 979.330220] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da5f9191-53b4-42a9-841f-466e20250114 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.346873] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 979.346873] env[61964]: value = "task-1041405" [ 979.346873] env[61964]: _type = "Task" [ 979.346873] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.357263] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041405, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.400521] env[61964]: DEBUG nova.compute.manager [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 979.403250] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.403250] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d9fa8e-b4f5-471b-94bf-b83e8df2695f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.413786] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 979.413864] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb16098f-4e10-44cd-bca0-7424155dac00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.420263] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 979.420263] env[61964]: value = "task-1041406" [ 979.420263] env[61964]: _type = "Task" [ 979.420263] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.429497] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041406, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.505143] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 979.546913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.552142] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.357s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.552467] env[61964]: DEBUG nova.objects.instance [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'resources' on Instance uuid 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.585620] env[61964]: INFO nova.scheduler.client.report [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance 4d9c24cf-e11b-4d95-ad41-538d91279689 [ 979.672812] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.729824] env[61964]: DEBUG nova.network.neutron [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updating instance_info_cache with network_info: [{"id": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "address": "fa:16:3e:03:af:6a", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0e7e755-ae", "ovs_interfaceid": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.856370] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041405, 'name': ReconfigVM_Task, 'duration_secs': 0.202084} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.856569] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 979.930309] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041406, 'name': PowerOffVM_Task, 'duration_secs': 0.156789} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.931973] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 979.931973] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 979.931973] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7ae36bd1-7fb5-40d3-9c63-5c9d3af1ccea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.992645] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 979.992887] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 979.993097] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleting the datastore file [datastore1] f5acabe4-674b-4c4c-85b4-b9270926017b {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.993382] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f06ab631-2413-4803-bb00-bd5f80a0749e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.000837] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for the task: (returnval){ [ 980.000837] env[61964]: value = "task-1041408" [ 980.000837] env[61964]: _type = "Task" [ 980.000837] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.008961] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041408, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.100595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-9c542469-d3dc-47fa-8a43-d16062784822 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "4d9c24cf-e11b-4d95-ad41-538d91279689" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.514s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.176666] env[61964]: DEBUG oslo_vmware.api [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041403, 'name': ReconfigVM_Task, 'duration_secs': 5.729567} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.179690] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.179956] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Reconfigured VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 980.223356] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfad3123-9c6d-436d-9ddc-9a53781c06db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.231485] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09bd24a-17e4-4e95-ad29-ada780a92931 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.234947] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.235312] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Instance network_info: |[{"id": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "address": "fa:16:3e:03:af:6a", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0e7e755-ae", "ovs_interfaceid": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 980.235567] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Acquired lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.235694] env[61964]: DEBUG nova.network.neutron [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Refreshing network info cache for port f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.237700] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:af:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97b68ed7-8461-4345-b064-96a1dde53a86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0e7e755-ae40-4d0d-bf57-2706d53067d6', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.245635] env[61964]: DEBUG oslo.service.loopingcall [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.248888] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 980.274645] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9a9cb44-2b2d-488b-a4c0-fef55dc9c029 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.290310] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b555b0b9-6903-4db1-8248-183b649fc47c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.300369] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b23f0c-4624-47d3-b47a-76df3974b398 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.304558] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.304558] env[61964]: value = "task-1041409" [ 980.304558] env[61964]: _type = "Task" [ 980.304558] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.318306] env[61964]: DEBUG nova.compute.provider_tree [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.326969] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041409, 'name': CreateVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.363385] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 980.363687] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.363910] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 980.364339] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.364435] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 980.364760] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 980.365092] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 980.365305] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 980.365566] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 980.365803] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 980.366073] env[61964]: DEBUG nova.virt.hardware [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 980.372246] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 980.372901] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf408025-be10-4fc8-85f0-28ec27f88b41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.394727] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 980.394727] env[61964]: value = "task-1041410" [ 980.394727] env[61964]: _type = "Task" [ 980.394727] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.402772] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041410, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.511114] env[61964]: DEBUG oslo_vmware.api [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Task: {'id': task-1041408, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140447} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.511393] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.511586] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 980.511775] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 980.512283] env[61964]: INFO nova.compute.manager [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 980.512283] env[61964]: DEBUG oslo.service.loopingcall [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.512475] env[61964]: DEBUG nova.compute.manager [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 980.512571] env[61964]: DEBUG nova.network.neutron [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 980.717497] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.717721] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.814352] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041409, 'name': CreateVM_Task, 'duration_secs': 0.45245} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.814579] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 980.815359] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.815608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.815987] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.816314] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba92548f-dcc6-47f4-bde0-cffc2fc4af7e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.823073] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 980.823073] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528977ce-3603-ac11-a371-af93e33914f4" [ 980.823073] env[61964]: _type = "Task" [ 980.823073] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.843145] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528977ce-3603-ac11-a371-af93e33914f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009252} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.843497] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.843785] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.844090] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.844304] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.844555] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.845085] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96671794-205e-400f-9ea6-ce96840b1fd2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.852178] env[61964]: ERROR nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [req-57a28610-b953-42b8-9f87-087a4dc63b51] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-57a28610-b953-42b8-9f87-087a4dc63b51"}]} [ 980.855773] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.856017] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 980.856757] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dd3dc1b-378c-4beb-9431-b284b1764e1b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.861775] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 980.861775] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526f3bf4-1f6f-d0b8-ce5c-c822897b7e6f" [ 980.861775] env[61964]: _type = "Task" [ 980.861775] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.870057] env[61964]: DEBUG nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 980.874877] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526f3bf4-1f6f-d0b8-ce5c-c822897b7e6f, 'name': SearchDatastore_Task} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.876142] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4eebf5e-0bd0-4c31-b88d-63b79b1328ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.881082] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 980.881082] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229b203-68dd-cdd9-6989-fc201fa4193a" [ 980.881082] env[61964]: _type = "Task" [ 980.881082] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.890239] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229b203-68dd-cdd9-6989-fc201fa4193a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.891172] env[61964]: DEBUG nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 980.891430] env[61964]: DEBUG nova.compute.provider_tree [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.904906] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041410, 'name': ReconfigVM_Task, 'duration_secs': 0.203101} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.905757] env[61964]: DEBUG nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 980.907612] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 980.908860] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e20ea9-b213-4db9-b706-7ddb2183875b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.930143] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.933328] env[61964]: DEBUG nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 980.935161] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4bd8744-bed4-4dc6-8706-889523e65792 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.953933] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 980.953933] env[61964]: value = "task-1041411" [ 980.953933] env[61964]: _type = "Task" [ 980.953933] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.961726] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.021456] env[61964]: DEBUG nova.network.neutron [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updated VIF entry in instance network info cache for port f0e7e755-ae40-4d0d-bf57-2706d53067d6. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 981.021882] env[61964]: DEBUG nova.network.neutron [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updating instance_info_cache with network_info: [{"id": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "address": "fa:16:3e:03:af:6a", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0e7e755-ae", "ovs_interfaceid": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.112884] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44202373-dcf5-4c71-b5bb-c91dba185229 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.120714] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23a4428-909e-4280-b1cd-0a2d22aeec54 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.151848] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a3de27-1c8c-4b3c-ac0b-ee7d2a624949 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.159959] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9679d17-fe7e-4f7e-b12b-42f06a7e1e50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.174471] env[61964]: DEBUG nova.compute.provider_tree [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.219933] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 981.225228] env[61964]: DEBUG nova.compute.manager [req-e84c9c08-5646-41fa-9b5c-8dd799e1aa9c req-23067268-6b14-49c2-8b74-4f12d3fa2190 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Received event network-vif-deleted-c820ead7-034a-450f-839a-9bca2af7bb43 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 981.225481] env[61964]: INFO nova.compute.manager [req-e84c9c08-5646-41fa-9b5c-8dd799e1aa9c req-23067268-6b14-49c2-8b74-4f12d3fa2190 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Neutron deleted interface c820ead7-034a-450f-839a-9bca2af7bb43; detaching it from the instance and deleting it from the info cache [ 981.225743] env[61964]: DEBUG nova.network.neutron [req-e84c9c08-5646-41fa-9b5c-8dd799e1aa9c req-23067268-6b14-49c2-8b74-4f12d3fa2190 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.337147] env[61964]: DEBUG nova.network.neutron [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.351574] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.351717] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.351860] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 981.391183] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229b203-68dd-cdd9-6989-fc201fa4193a, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.391456] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.391711] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 90679b1f-2c9c-4049-bfb9-db573220163b/90679b1f-2c9c-4049-bfb9-db573220163b.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 981.391973] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf9da8a2-64cc-4c5b-964e-ca2f4d43631c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.397857] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 981.397857] env[61964]: value = "task-1041412" [ 981.397857] env[61964]: _type = "Task" [ 981.397857] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.405860] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.463884] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041411, 'name': ReconfigVM_Task, 'duration_secs': 0.322295} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.464194] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 39683f54-fa99-448d-91ab-171397be48aa/39683f54-fa99-448d-91ab-171397be48aa.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.464476] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 981.525347] env[61964]: DEBUG oslo_concurrency.lockutils [req-6229ecdc-7ffe-4d32-8f12-862b79b17f54 req-fd7a1792-494d-459c-8400-3224746ae5b3 service nova] Releasing lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.587534] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.587740] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.587983] env[61964]: DEBUG nova.network.neutron [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 981.707733] env[61964]: DEBUG nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 122 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 981.707974] env[61964]: DEBUG nova.compute.provider_tree [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 122 to 123 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 981.708265] env[61964]: DEBUG nova.compute.provider_tree [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.734366] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b53574a-a182-4896-9919-67b15b3e9275 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.745565] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e4b815-73bf-4825-a50a-4fdaf6d8d34b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.759763] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.778992] env[61964]: DEBUG nova.compute.manager [req-e84c9c08-5646-41fa-9b5c-8dd799e1aa9c req-23067268-6b14-49c2-8b74-4f12d3fa2190 service nova] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Detach interface failed, port_id=c820ead7-034a-450f-839a-9bca2af7bb43, reason: Instance f5acabe4-674b-4c4c-85b4-b9270926017b could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 981.840089] env[61964]: INFO nova.compute.manager [-] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Took 1.33 seconds to deallocate network for instance. [ 981.907718] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454735} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.908258] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 90679b1f-2c9c-4049-bfb9-db573220163b/90679b1f-2c9c-4049-bfb9-db573220163b.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 981.908493] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.908749] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c199a192-ecc1-4bee-a90b-1f790b8b69d1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.917171] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 981.917171] env[61964]: value = "task-1041413" [ 981.917171] env[61964]: _type = "Task" [ 981.917171] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.923689] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041413, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.971311] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535bdd3a-4d9b-4953-9ec8-afa1412438c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.991274] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2c6e25-a155-4b26-85e3-44d78e594669 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.008353] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 982.213586] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.661s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.218038] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.548s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.218038] env[61964]: DEBUG nova.objects.instance [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'resources' on Instance uuid f983316e-e012-41fd-bf20-7d9cc5accd71 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.239508] env[61964]: INFO nova.scheduler.client.report [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleted allocations for instance 138f44e4-e12e-4f89-a9b2-8a512b53cdf5 [ 982.346153] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.425324] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041413, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069544} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.425893] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.426650] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e640b4-6d67-414e-b602-95a6410a9cfc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.447520] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 90679b1f-2c9c-4049-bfb9-db573220163b/90679b1f-2c9c-4049-bfb9-db573220163b.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.447765] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a36fc82-3eba-4b92-96ac-fe093ea96998 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.471678] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 982.471678] env[61964]: value = "task-1041414" [ 982.471678] env[61964]: _type = "Task" [ 982.471678] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.479101] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.554046] env[61964]: DEBUG nova.network.neutron [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Port 79400d4a-e480-4d60-8c9f-dcb1b899ae7c binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 982.664063] env[61964]: INFO nova.network.neutron [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Port e678489f-0505-421c-893c-43523e7e29f7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 982.664446] env[61964]: DEBUG nova.network.neutron [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.754106] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a63ec44d-74dc-4eab-a83a-a7a929d66c20 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "138f44e4-e12e-4f89-a9b2-8a512b53cdf5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.843s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.791479] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updating instance_info_cache with network_info: [{"id": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "address": "fa:16:3e:32:c5:75", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74f7cab9-da", "ovs_interfaceid": "74f7cab9-da08-4246-bebd-f8b2f0ec5a3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.941451] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4587247d-c414-4a8f-9129-153990448c63 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.949717] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18a84ac-5617-4696-b212-9586d3bbdfd6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.996153] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8670f45c-cc56-4cc1-9173-1ca3e1ef93ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.004180] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.007439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0aaf764-17a5-43ae-b20a-a8e5fbef8622 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.021953] env[61964]: DEBUG nova.compute.provider_tree [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 983.170882] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.257172] env[61964]: DEBUG nova.compute.manager [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 983.258331] env[61964]: DEBUG nova.compute.manager [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing instance network info cache due to event network-changed-083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 983.258331] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Acquiring lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.258331] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Acquired lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.258331] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Refreshing network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 983.296035] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-aef4c3a7-641a-4356-9187-ae4c082ccde9" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.296035] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 983.296035] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296177] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296238] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296450] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296514] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296658] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.296830] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 983.296986] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.501163] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041414, 'name': ReconfigVM_Task, 'duration_secs': 0.596065} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.501577] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 90679b1f-2c9c-4049-bfb9-db573220163b/90679b1f-2c9c-4049-bfb9-db573220163b.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.502087] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e0b4f80-3705-4226-84d1-69f72aaac87e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.508416] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 983.508416] env[61964]: value = "task-1041415" [ 983.508416] env[61964]: _type = "Task" [ 983.508416] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.516235] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041415, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.553898] env[61964]: DEBUG nova.scheduler.client.report [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 983.554214] env[61964]: DEBUG nova.compute.provider_tree [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 123 to 124 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 983.554412] env[61964]: DEBUG nova.compute.provider_tree [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 983.575974] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.576307] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.576507] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.617617] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.618602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.618602] env[61964]: DEBUG nova.objects.instance [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'flavor' on Instance uuid c4fabe85-1b95-4981-9e05-dbfedbbe1e1e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.675204] env[61964]: DEBUG oslo_concurrency.lockutils [None req-38f9da20-08b0-4746-a12f-dd886dd6b77b tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-47378856-4cbf-4cf6-aecd-ae935885df3d-e678489f-0505-421c-893c-43523e7e29f7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.104s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.802564] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.960261] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updated VIF entry in instance network info cache for port 083961bb-ae89-4d60-8ea8-3d5454613407. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 983.960507] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [{"id": "083961bb-ae89-4d60-8ea8-3d5454613407", "address": "fa:16:3e:ac:3b:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap083961bb-ae", "ovs_interfaceid": "083961bb-ae89-4d60-8ea8-3d5454613407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.021522] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041415, 'name': Rename_Task, 'duration_secs': 0.129697} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.021761] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 984.022016] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7e8e6a4-aac5-4ba8-9355-b807e6886bd0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.028424] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 984.028424] env[61964]: value = "task-1041416" [ 984.028424] env[61964]: _type = "Task" [ 984.028424] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.037640] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.059731] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.062047] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.288s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.063555] env[61964]: INFO nova.compute.claims [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.085163] env[61964]: INFO nova.scheduler.client.report [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance f983316e-e012-41fd-bf20-7d9cc5accd71 [ 984.227178] env[61964]: DEBUG nova.objects.instance [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'pci_requests' on Instance uuid c4fabe85-1b95-4981-9e05-dbfedbbe1e1e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.463455] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Releasing lock "refresh_cache-47378856-4cbf-4cf6-aecd-ae935885df3d" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.463792] env[61964]: DEBUG nova.compute.manager [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 984.464016] env[61964]: DEBUG nova.compute.manager [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing instance network info cache due to event network-changed-bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 984.464277] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.464456] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.464656] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.539299] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041416, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.597902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-fb37319d-cfd6-4cf5-990d-15f9eb40df00 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f983316e-e012-41fd-bf20-7d9cc5accd71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.209s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.611454] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.611643] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.611832] env[61964]: DEBUG nova.network.neutron [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 984.729613] env[61964]: DEBUG nova.objects.base [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 984.729851] env[61964]: DEBUG nova.network.neutron [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 984.760483] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.760714] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.797362] env[61964]: DEBUG nova.policy [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '076bc383863540faabab1f463a01f7dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d4175848e8e492ba4f749e464b693ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.038332] env[61964]: DEBUG oslo_vmware.api [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041416, 'name': PowerOnVM_Task, 'duration_secs': 0.630182} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.038829] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 985.039105] env[61964]: INFO nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Took 7.17 seconds to spawn the instance on the hypervisor. [ 985.039342] env[61964]: DEBUG nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 985.040130] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d3bba8-94a5-4741-9be0-3471d46f5e1a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.222603] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updated VIF entry in instance network info cache for port bc9cb29f-849e-4bea-8532-cd8b9efcb198. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.222978] env[61964]: DEBUG nova.network.neutron [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.245550] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753eba0f-b768-4df9-a3a7-5866da66dffa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.253698] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bcfa2e-4331-4120-9164-a883df3ca6a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.288625] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 985.294390] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a75437-ab3f-4ff5-9619-8d6d0e862754 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.301899] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad3f2a6-8502-40f7-93b7-604840f59145 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.314944] env[61964]: DEBUG nova.compute.provider_tree [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.349265] env[61964]: DEBUG nova.network.neutron [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.456935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.457307] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.558024] env[61964]: INFO nova.compute.manager [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Took 16.40 seconds to build instance. [ 985.727899] env[61964]: DEBUG oslo_concurrency.lockutils [req-65adbe37-54bc-4a16-a620-2966beb68fc0 req-22934859-11d5-4efd-b42e-3b2d3a7af12b service nova] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.806417] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.822049] env[61964]: DEBUG nova.scheduler.client.report [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 985.852851] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.962021] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 986.059846] env[61964]: DEBUG oslo_concurrency.lockutils [None req-88d1d2eb-a7b8-4659-bd46-ab4c9cccae09 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.906s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.236817] env[61964]: DEBUG nova.compute.manager [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Received event network-changed-f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 986.237068] env[61964]: DEBUG nova.compute.manager [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Refreshing instance network info cache due to event network-changed-f0e7e755-ae40-4d0d-bf57-2706d53067d6. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 986.237315] env[61964]: DEBUG oslo_concurrency.lockutils [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] Acquiring lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.237464] env[61964]: DEBUG oslo_concurrency.lockutils [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] Acquired lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.237629] env[61964]: DEBUG nova.network.neutron [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Refreshing network info cache for port f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.324741] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.326076] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 986.328065] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.568s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.329502] env[61964]: INFO nova.compute.claims [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.374751] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e699de7f-41cb-42bc-9cf8-9f89b1d3d690 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.393903] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542740e7-f72c-4b90-add7-e0bfa9a7730d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.400684] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.481703] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.835528] env[61964]: DEBUG nova.compute.utils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.838813] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 986.838983] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 986.851125] env[61964]: DEBUG nova.network.neutron [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Successfully updated port: e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.883632] env[61964]: DEBUG nova.policy [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd63c376b2814a4ab07fea57284525a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8e993c6c7ed4548ae4e2fdafe92cb5b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.909868] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.911015] env[61964]: DEBUG nova.compute.manager [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 986.915615] env[61964]: DEBUG oslo_concurrency.lockutils [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.915615] env[61964]: DEBUG oslo_concurrency.lockutils [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.915615] env[61964]: DEBUG oslo_concurrency.lockutils [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.915615] env[61964]: DEBUG nova.compute.manager [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] No waiting events found dispatching network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 986.915615] env[61964]: WARNING nova.compute.manager [req-d31bc7d0-f705-4930-b839-08cc15c97fce req-c6de0366-92c4-43aa-bd35-c57648cd84bd service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received unexpected event network-vif-plugged-e678489f-0505-421c-893c-43523e7e29f7 for instance with vm_state active and task_state None. [ 986.916569] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69156c4d-68a7-458c-be95-6f055023b718 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.924750] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 986.924750] env[61964]: value = "task-1041417" [ 986.924750] env[61964]: _type = "Task" [ 986.924750] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.935612] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041417, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.043931] env[61964]: DEBUG nova.network.neutron [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updated VIF entry in instance network info cache for port f0e7e755-ae40-4d0d-bf57-2706d53067d6. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.044434] env[61964]: DEBUG nova.network.neutron [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updating instance_info_cache with network_info: [{"id": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "address": "fa:16:3e:03:af:6a", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0e7e755-ae", "ovs_interfaceid": "f0e7e755-ae40-4d0d-bf57-2706d53067d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.148174] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Successfully created port: 1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.340437] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 987.354018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.354431] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.354629] env[61964]: DEBUG nova.network.neutron [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.439403] env[61964]: DEBUG oslo_vmware.api [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041417, 'name': PowerOnVM_Task, 'duration_secs': 0.495772} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.439696] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.439890] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bd89ee4f-1659-483a-a35c-7b5d6c01bcec tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance '39683f54-fa99-448d-91ab-171397be48aa' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.540277] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5a3e59-7bf0-4bab-b250-581341d6c2d2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.548068] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e6177d-398d-49d0-ae2e-3544e12fee75 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.551342] env[61964]: DEBUG oslo_concurrency.lockutils [req-93aeff9e-3aa4-4a23-b495-8fca7460267d req-90bc0076-7f45-4713-90c5-22efc9cd3b36 service nova] Releasing lock "refresh_cache-90679b1f-2c9c-4049-bfb9-db573220163b" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.578341] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64596bcf-531f-43fd-a4e9-173acabe67be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.586019] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0382fd-5b12-494d-8ae8-806f4fd5582a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.599470] env[61964]: DEBUG nova.compute.provider_tree [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.901692] env[61964]: WARNING nova.network.neutron [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] 48dd64e7-3659-4104-8273-6f58f8d3eb44 already exists in list: networks containing: ['48dd64e7-3659-4104-8273-6f58f8d3eb44']. ignoring it [ 988.102420] env[61964]: DEBUG nova.scheduler.client.report [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 988.255897] env[61964]: DEBUG nova.network.neutron [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e678489f-0505-421c-893c-43523e7e29f7", "address": "fa:16:3e:3a:ec:64", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape678489f-05", "ovs_interfaceid": "e678489f-0505-421c-893c-43523e7e29f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.350655] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 988.376028] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 988.376305] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.376470] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 988.376657] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.376810] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 988.376962] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 988.377602] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 988.377602] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 988.377602] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 988.377743] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 988.377926] env[61964]: DEBUG nova.virt.hardware [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 988.378846] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129f8160-c4a4-4182-9fc7-e0a3c0b70889 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.386827] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3fa1d0-664a-408f-91e6-7fb4e3551789 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.552077] env[61964]: DEBUG nova.compute.manager [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Received event network-vif-plugged-1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 988.552380] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] Acquiring lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.552555] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.552728] env[61964]: DEBUG oslo_concurrency.lockutils [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.552902] env[61964]: DEBUG nova.compute.manager [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] No waiting events found dispatching network-vif-plugged-1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 988.553261] env[61964]: WARNING nova.compute.manager [req-4b26e42f-038a-4c51-8d9a-88d3081a234d req-ad83a21c-4077-4c66-92bb-595494b3c5c4 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Received unexpected event network-vif-plugged-1081d12c-3c4a-471e-b29c-c96b1bfed882 for instance with vm_state building and task_state spawning. [ 988.607679] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.608291] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 988.610955] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.265s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.611204] env[61964]: DEBUG nova.objects.instance [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lazy-loading 'resources' on Instance uuid f5acabe4-674b-4c4c-85b4-b9270926017b {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.759140] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.759822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.759993] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.760851] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5b617b-a7b4-4399-b55d-0b2f45605b46 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.778014] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 988.778242] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.778410] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 988.778621] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.778751] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 988.778899] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 988.779132] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 988.779321] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 988.779530] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 988.779706] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 988.779885] env[61964]: DEBUG nova.virt.hardware [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 988.785994] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfiguring VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 988.786579] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-384bbede-8eb6-428c-9c36-fdbd7f0296c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.803724] env[61964]: DEBUG oslo_vmware.api [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 988.803724] env[61964]: value = "task-1041418" [ 988.803724] env[61964]: _type = "Task" [ 988.803724] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.811121] env[61964]: DEBUG oslo_vmware.api [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041418, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.931633] env[61964]: DEBUG nova.compute.manager [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-changed-e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 988.931881] env[61964]: DEBUG nova.compute.manager [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing instance network info cache due to event network-changed-e678489f-0505-421c-893c-43523e7e29f7. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 988.932059] env[61964]: DEBUG oslo_concurrency.lockutils [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.932211] env[61964]: DEBUG oslo_concurrency.lockutils [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.932375] env[61964]: DEBUG nova.network.neutron [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Refreshing network info cache for port e678489f-0505-421c-893c-43523e7e29f7 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.106548] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Successfully updated port: 1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.114497] env[61964]: DEBUG nova.compute.utils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.119033] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 989.119263] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.179157] env[61964]: DEBUG nova.policy [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.298541] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c094e2-1177-48d4-a435-dd457621fcfb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.309248] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227cf04a-ffcd-4cb3-b826-6fa3502eb93f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.317110] env[61964]: DEBUG oslo_vmware.api [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.341746] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45179a51-76fa-451a-b792-c0807475277e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.349218] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf0c64f-7cd7-473e-b3f7-cf94ab045db6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.362625] env[61964]: DEBUG nova.compute.provider_tree [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.468830] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Successfully created port: 4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.612396] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.616029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquired lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.616029] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.620140] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 989.739337] env[61964]: DEBUG nova.network.neutron [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updated VIF entry in instance network info cache for port e678489f-0505-421c-893c-43523e7e29f7. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.739965] env[61964]: DEBUG nova.network.neutron [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e678489f-0505-421c-893c-43523e7e29f7", "address": "fa:16:3e:3a:ec:64", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape678489f-05", "ovs_interfaceid": "e678489f-0505-421c-893c-43523e7e29f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.770890] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.771183] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.771377] env[61964]: DEBUG nova.compute.manager [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Going to confirm migration 4 {{(pid=61964) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5218}} [ 989.821854] env[61964]: DEBUG oslo_vmware.api [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041418, 'name': ReconfigVM_Task, 'duration_secs': 0.774618} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.822376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.822606] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfigured VM to attach interface {{(pid=61964) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 989.865867] env[61964]: DEBUG nova.scheduler.client.report [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.154073] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 990.242719] env[61964]: DEBUG oslo_concurrency.lockutils [req-e13a987d-7a09-493c-992e-153b54b0c5ba req-a5f386d3-828f-4b41-98f8-de2b3c3428e6 service nova] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.327238] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df0d5b39-5a69-4d86-8d61-b4ca57abf421 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.709s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.329068] env[61964]: DEBUG nova.network.neutron [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Updating instance_info_cache with network_info: [{"id": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "address": "fa:16:3e:84:6f:cc", "network": {"id": "ce871e12-2dbd-4d6d-a515-9e40d333427c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-489470647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8e993c6c7ed4548ae4e2fdafe92cb5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1081d12c-3c", "ovs_interfaceid": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.336058] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.336258] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.336426] env[61964]: DEBUG nova.network.neutron [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.336676] env[61964]: DEBUG nova.objects.instance [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'info_cache' on Instance uuid 39683f54-fa99-448d-91ab-171397be48aa {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.372173] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.377916] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.575s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.382988] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.382988] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 990.382988] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.575s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.382988] env[61964]: INFO nova.compute.claims [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.386965] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35317a1a-1167-4741-a2a9-387b4e2c1763 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.400218] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b3e2bb-edef-4787-9c40-6d418996f1c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.416048] env[61964]: INFO nova.scheduler.client.report [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Deleted allocations for instance f5acabe4-674b-4c4c-85b4-b9270926017b [ 990.421571] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd2a1c1-0178-4e49-a94a-444d60a22eae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.430596] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9f60e2-0ab4-40c7-9003-4c2a4dd74f88 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.464667] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179983MB free_disk=182GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 990.464862] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.630765] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 990.656234] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 990.656494] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.656660] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 990.656846] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.656998] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 990.657219] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 990.657457] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 990.657623] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 990.657793] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 990.657959] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 990.658208] env[61964]: DEBUG nova.virt.hardware [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 990.659089] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16bb50b8-dac1-4152-a37c-e095357ab977 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.666634] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76b6a46-87a6-4d02-aadd-58f2ff530c6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.832391] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Releasing lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.832703] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Instance network_info: |[{"id": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "address": "fa:16:3e:84:6f:cc", "network": {"id": "ce871e12-2dbd-4d6d-a515-9e40d333427c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-489470647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8e993c6c7ed4548ae4e2fdafe92cb5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1081d12c-3c", "ovs_interfaceid": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 990.833147] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:6f:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '678ebbe4-4c53-4eaf-a689-93981310f37d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1081d12c-3c4a-471e-b29c-c96b1bfed882', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.841413] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Creating folder: Project (f8e993c6c7ed4548ae4e2fdafe92cb5b). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 990.844159] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8ed9997-105d-4391-a7ab-1ac29f400579 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.855820] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Created folder: Project (f8e993c6c7ed4548ae4e2fdafe92cb5b) in parent group-v230360. [ 990.856069] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Creating folder: Instances. Parent ref: group-v230549. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 990.856355] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5dc8bc67-d61e-479f-a06d-f8377a2309e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.866101] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Created folder: Instances in parent group-v230549. [ 990.866391] env[61964]: DEBUG oslo.service.loopingcall [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.866594] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 990.866799] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7a25e4d-32d6-44f9-af1d-f9642c71f109 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.888479] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.888479] env[61964]: value = "task-1041421" [ 990.888479] env[61964]: _type = "Task" [ 990.888479] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.899213] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041421, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.923697] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22835360-652f-4272-917e-c7eb5adb3c41 tempest-ListServerFiltersTestJSON-1728165309 tempest-ListServerFiltersTestJSON-1728165309-project-member] Lock "f5acabe4-674b-4c4c-85b4-b9270926017b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.030s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.961725] env[61964]: DEBUG nova.compute.manager [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Received event network-changed-1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 990.961930] env[61964]: DEBUG nova.compute.manager [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Refreshing instance network info cache due to event network-changed-1081d12c-3c4a-471e-b29c-c96b1bfed882. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 990.962169] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Acquiring lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.962399] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Acquired lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.962571] env[61964]: DEBUG nova.network.neutron [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Refreshing network info cache for port 1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 990.972290] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Successfully updated port: 4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.403514] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041421, 'name': CreateVM_Task, 'duration_secs': 0.343223} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.403877] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 991.404548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.404725] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.409623] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 991.409886] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-227e432b-2f1d-4728-8c9f-240ed4792dc4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.415085] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 991.415085] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f2f93-6b95-aeca-0922-59a1465682eb" [ 991.415085] env[61964]: _type = "Task" [ 991.415085] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.426337] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f2f93-6b95-aeca-0922-59a1465682eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.474988] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.475164] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.475343] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.596723] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062ef4cf-9747-4143-8050-7b7250460d98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.607339] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7572c923-2703-4962-81c4-1f49123f2d90 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.644401] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da1369c-0e1c-4566-8351-8603ee754e6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.652702] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c042f24-33cf-48ef-bb58-6a7c3d485272 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.667290] env[61964]: DEBUG nova.compute.provider_tree [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.729360] env[61964]: DEBUG nova.network.neutron [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [{"id": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "address": "fa:16:3e:cb:01:66", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79400d4a-e4", "ovs_interfaceid": "79400d4a-e480-4d60-8c9f-dcb1b899ae7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.805799] env[61964]: DEBUG nova.network.neutron [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Updated VIF entry in instance network info cache for port 1081d12c-3c4a-471e-b29c-c96b1bfed882. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 991.806189] env[61964]: DEBUG nova.network.neutron [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Updating instance_info_cache with network_info: [{"id": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "address": "fa:16:3e:84:6f:cc", "network": {"id": "ce871e12-2dbd-4d6d-a515-9e40d333427c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-489470647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8e993c6c7ed4548ae4e2fdafe92cb5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "678ebbe4-4c53-4eaf-a689-93981310f37d", "external-id": "nsx-vlan-transportzone-443", "segmentation_id": 443, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1081d12c-3c", "ovs_interfaceid": "1081d12c-3c4a-471e-b29c-c96b1bfed882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.850846] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.851126] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.926703] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527f2f93-6b95-aeca-0922-59a1465682eb, 'name': SearchDatastore_Task, 'duration_secs': 0.010734} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.926703] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.926878] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.927061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.927279] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.927493] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.927746] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96acb626-7e57-4c7e-b20b-6a99633348af {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.935379] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.935614] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 991.936481] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc2c5fa6-270e-422a-a3b1-8673cf27b877 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.940927] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 991.940927] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c2f1d1-efe2-7000-a361-6b6f0aef7815" [ 991.940927] env[61964]: _type = "Task" [ 991.940927] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.949607] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c2f1d1-efe2-7000-a361-6b6f0aef7815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.006975] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.123800] env[61964]: DEBUG nova.network.neutron [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Updating instance_info_cache with network_info: [{"id": "4d6064e2-9257-47a1-93d4-5208c08288f2", "address": "fa:16:3e:ae:fd:b3", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d6064e2-92", "ovs_interfaceid": "4d6064e2-9257-47a1-93d4-5208c08288f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.170795] env[61964]: DEBUG nova.scheduler.client.report [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.232351] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-39683f54-fa99-448d-91ab-171397be48aa" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.232601] env[61964]: DEBUG nova.objects.instance [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'migration_context' on Instance uuid 39683f54-fa99-448d-91ab-171397be48aa {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.309679] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Releasing lock "refresh_cache-2436ae21-6244-4f8b-abb0-1e7530f07dc5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.309949] env[61964]: DEBUG nova.compute.manager [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Received event network-vif-plugged-4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 992.310168] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Acquiring lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.310381] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.310548] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.310717] env[61964]: DEBUG nova.compute.manager [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] No waiting events found dispatching network-vif-plugged-4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 992.310888] env[61964]: WARNING nova.compute.manager [req-dc016734-29fe-4dc4-afbd-3424a80e20a4 req-0f1eff43-9055-4208-8107-efb1663aa895 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Received unexpected event network-vif-plugged-4d6064e2-9257-47a1-93d4-5208c08288f2 for instance with vm_state building and task_state spawning. [ 992.353986] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.354246] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.355195] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4faddc-0578-4664-97da-6b01c61f596f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.374014] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dbe2d6-6bde-4ad1-9c3b-5862293663f9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.400173] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfiguring VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 992.400479] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0a6fcb8-2a31-4034-959a-8e1736afcfda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.419188] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 992.419188] env[61964]: value = "task-1041422" [ 992.419188] env[61964]: _type = "Task" [ 992.419188] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.427572] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.451948] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c2f1d1-efe2-7000-a361-6b6f0aef7815, 'name': SearchDatastore_Task, 'duration_secs': 0.008958} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.453070] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-963e86b7-4b2f-41d7-953b-812d450ec0ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.459165] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 992.459165] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522add3e-ab1a-59fd-bc30-9da9f4d22358" [ 992.459165] env[61964]: _type = "Task" [ 992.459165] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.467462] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522add3e-ab1a-59fd-bc30-9da9f4d22358, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.627055] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.627444] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Instance network_info: |[{"id": "4d6064e2-9257-47a1-93d4-5208c08288f2", "address": "fa:16:3e:ae:fd:b3", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d6064e2-92", "ovs_interfaceid": "4d6064e2-9257-47a1-93d4-5208c08288f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 992.627904] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:fd:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d6064e2-9257-47a1-93d4-5208c08288f2', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.635313] env[61964]: DEBUG oslo.service.loopingcall [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.635545] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 992.636116] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1c07468-e0b1-484a-b187-6a3ffe5940ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.656958] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.656958] env[61964]: value = "task-1041423" [ 992.656958] env[61964]: _type = "Task" [ 992.656958] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.664644] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041423, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.675949] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.676252] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 992.678939] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.197s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.681017] env[61964]: INFO nova.compute.claims [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.735504] env[61964]: DEBUG nova.objects.base [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Object Instance<39683f54-fa99-448d-91ab-171397be48aa> lazy-loaded attributes: info_cache,migration_context {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 992.736368] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff15f0e-cbe8-4121-ba22-89e65927927e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.756931] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beccfa0b-240e-4f6a-9ce1-afea7bed59d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.762135] env[61964]: DEBUG oslo_vmware.api [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 992.762135] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cd85b0-f454-3788-fc5e-d34c94c768e6" [ 992.762135] env[61964]: _type = "Task" [ 992.762135] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.769676] env[61964]: DEBUG oslo_vmware.api [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cd85b0-f454-3788-fc5e-d34c94c768e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.929061] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.969475] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522add3e-ab1a-59fd-bc30-9da9f4d22358, 'name': SearchDatastore_Task, 'duration_secs': 0.009101} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.969770] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.970050] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 2436ae21-6244-4f8b-abb0-1e7530f07dc5/2436ae21-6244-4f8b-abb0-1e7530f07dc5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 992.970316] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36d9b278-de89-4516-8f48-a232ed961fca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.977122] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 992.977122] env[61964]: value = "task-1041424" [ 992.977122] env[61964]: _type = "Task" [ 992.977122] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.985129] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.987207] env[61964]: DEBUG nova.compute.manager [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Received event network-changed-4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 992.987400] env[61964]: DEBUG nova.compute.manager [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Refreshing instance network info cache due to event network-changed-4d6064e2-9257-47a1-93d4-5208c08288f2. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 992.987646] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] Acquiring lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.987849] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] Acquired lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.988046] env[61964]: DEBUG nova.network.neutron [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Refreshing network info cache for port 4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.168367] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041423, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.185084] env[61964]: DEBUG nova.compute.utils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.190114] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 993.190114] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 993.232518] env[61964]: DEBUG nova.policy [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c29d89b14fc4a12bf6b35e94b31373e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6110872ae3dc4491bb10abb1945ffe2a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.274927] env[61964]: DEBUG oslo_vmware.api [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cd85b0-f454-3788-fc5e-d34c94c768e6, 'name': SearchDatastore_Task, 'duration_secs': 0.007455} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.275353] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.429527] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.486666] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43724} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.486927] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 2436ae21-6244-4f8b-abb0-1e7530f07dc5/2436ae21-6244-4f8b-abb0-1e7530f07dc5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 993.487162] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.487440] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c76e9f4-8a15-4c43-bd9a-e0a222f44974 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.494180] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 993.494180] env[61964]: value = "task-1041425" [ 993.494180] env[61964]: _type = "Task" [ 993.494180] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.501039] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.593479] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Successfully created port: ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.669700] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041423, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.689690] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 993.873930] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c485775f-43fd-4d8a-81e6-29d741b46991 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.881648] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf0b550-adcf-4a5f-9f28-47ad46bb6c1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.911800] env[61964]: DEBUG nova.network.neutron [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Updated VIF entry in instance network info cache for port 4d6064e2-9257-47a1-93d4-5208c08288f2. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.912220] env[61964]: DEBUG nova.network.neutron [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Updating instance_info_cache with network_info: [{"id": "4d6064e2-9257-47a1-93d4-5208c08288f2", "address": "fa:16:3e:ae:fd:b3", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d6064e2-92", "ovs_interfaceid": "4d6064e2-9257-47a1-93d4-5208c08288f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.913930] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9138b49b-de02-414b-946f-d3e44adeaa31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.926188] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94605a2d-4240-49c9-b582-af69a27b55f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.934983] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.944067] env[61964]: DEBUG nova.compute.provider_tree [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.003849] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062602} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.006773] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 994.006773] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dfdf36-9697-4792-b160-7d7887a3e335 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.028066] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 2436ae21-6244-4f8b-abb0-1e7530f07dc5/2436ae21-6244-4f8b-abb0-1e7530f07dc5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.028534] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea548327-95a5-46e5-9d16-135ce47084d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.047939] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 994.047939] env[61964]: value = "task-1041426" [ 994.047939] env[61964]: _type = "Task" [ 994.047939] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.055123] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.170704] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041423, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.418035] env[61964]: DEBUG oslo_concurrency.lockutils [req-b0692307-7b9e-4521-bdb5-d7d880a3a5fc req-1d3b13f5-d5fa-4b12-a933-62d0940dad49 service nova] Releasing lock "refresh_cache-81f821a3-6238-4ab1-9fe3-56d3f83ad52c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.430140] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.446113] env[61964]: DEBUG nova.scheduler.client.report [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 994.557316] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041426, 'name': ReconfigVM_Task, 'duration_secs': 0.345736} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.557636] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 2436ae21-6244-4f8b-abb0-1e7530f07dc5/2436ae21-6244-4f8b-abb0-1e7530f07dc5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.558323] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58e7bbdc-7ee0-4eff-8dd2-cef31f47eb3f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.563571] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 994.563571] env[61964]: value = "task-1041427" [ 994.563571] env[61964]: _type = "Task" [ 994.563571] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.571210] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041427, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.669141] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041423, 'name': CreateVM_Task, 'duration_secs': 1.550912} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.669317] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 994.670088] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.670278] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.670609] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.670879] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5a197d7-828a-4326-a79a-5a1ea78379d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.675335] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 994.675335] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ea632b-e7ec-036a-3b52-669af79a35be" [ 994.675335] env[61964]: _type = "Task" [ 994.675335] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.684139] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ea632b-e7ec-036a-3b52-669af79a35be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.703460] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 994.732425] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 994.732691] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.732873] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 994.733078] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.733236] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 994.733393] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 994.733606] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 994.733771] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 994.734702] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 994.734702] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 994.734702] env[61964]: DEBUG nova.virt.hardware [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 994.735154] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62536fe2-5a56-45c2-8d3a-2e611fabc884 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.743294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca2ba8c-5e33-4972-89df-0df5544db125 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.930631] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.951723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.952322] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 994.955086] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.490s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.074050] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041427, 'name': Rename_Task, 'duration_secs': 0.150472} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.074050] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.074326] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-549a0785-0200-4ff2-b121-e6abb24584e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.080597] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 995.080597] env[61964]: value = "task-1041428" [ 995.080597] env[61964]: _type = "Task" [ 995.080597] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.089866] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041428, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.185733] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52ea632b-e7ec-036a-3b52-669af79a35be, 'name': SearchDatastore_Task, 'duration_secs': 0.011255} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.186062] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.186309] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.186551] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.186704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.186886] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.187182] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ab2cbb7-cb10-49ba-af6c-d947611a94f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.194924] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.195113] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.195913] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f44f407b-d8b0-4d42-9cbf-6e2273bd48b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.200639] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 995.200639] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529e7573-7e52-bb32-2e9b-dcd310460e33" [ 995.200639] env[61964]: _type = "Task" [ 995.200639] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.208055] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529e7573-7e52-bb32-2e9b-dcd310460e33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.433087] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.441082] env[61964]: DEBUG nova.compute.manager [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Received event network-vif-plugged-ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 995.441082] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.441200] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.441347] env[61964]: DEBUG oslo_concurrency.lockutils [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.441519] env[61964]: DEBUG nova.compute.manager [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] No waiting events found dispatching network-vif-plugged-ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 995.441627] env[61964]: WARNING nova.compute.manager [req-2c85f221-2ad6-418a-a51b-df3493a74eea req-36246ffd-09af-4623-a2df-a1ad86fd22ec service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Received unexpected event network-vif-plugged-ed20338e-f092-4780-aede-3f9c39a0b9ea for instance with vm_state building and task_state spawning. [ 995.458474] env[61964]: DEBUG nova.compute.utils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.469236] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 995.469236] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 995.473442] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 995.544886] env[61964]: DEBUG nova.policy [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.579671] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Successfully updated port: ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.590331] env[61964]: DEBUG oslo_vmware.api [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041428, 'name': PowerOnVM_Task, 'duration_secs': 0.476317} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.591119] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 995.591339] env[61964]: INFO nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Took 7.24 seconds to spawn the instance on the hypervisor. [ 995.591528] env[61964]: DEBUG nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.592327] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ea0db7-211e-4bcb-8163-b6c993531ba3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.710542] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529e7573-7e52-bb32-2e9b-dcd310460e33, 'name': SearchDatastore_Task, 'duration_secs': 0.008363} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.711332] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f76bc98-9f93-4cee-bdc8-4e08c7976255 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.716523] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 995.716523] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522826ed-1321-dcf3-9826-f11852651b6b" [ 995.716523] env[61964]: _type = "Task" [ 995.716523] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.723847] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522826ed-1321-dcf3-9826-f11852651b6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.859849] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Successfully created port: 453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.933464] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.973148] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Applying migration context for instance 39683f54-fa99-448d-91ab-171397be48aa as it has an incoming, in-progress migration 9a2fc428-7e53-4418-89a7-f123879cb767. Migration status is confirming {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 995.974831] env[61964]: INFO nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating resource usage from migration 9a2fc428-7e53-4418-89a7-f123879cb767 [ 996.007283] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance aef4c3a7-641a-4356-9187-ae4c082ccde9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.007462] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 47378856-4cbf-4cf6-aecd-ae935885df3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.007629] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.007833] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance c4fabe85-1b95-4981-9e05-dbfedbbe1e1e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.007900] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 15e92512-3c92-43f0-891c-84bdaad1d186 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.007990] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Migration 9a2fc428-7e53-4418-89a7-f123879cb767 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 996.008121] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 39683f54-fa99-448d-91ab-171397be48aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008263] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 90679b1f-2c9c-4049-bfb9-db573220163b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008457] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 2436ae21-6244-4f8b-abb0-1e7530f07dc5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008566] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 81f821a3-6238-4ab1-9fe3-56d3f83ad52c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008627] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 0da26d16-84f2-4c00-97f5-ff132fa122d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008742] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance f7f18f7b-5e58-426c-aa5d-7063a50c32ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.008969] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 996.009125] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 996.085743] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.085894] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.086059] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.114280] env[61964]: INFO nova.compute.manager [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Took 18.36 seconds to build instance. [ 996.190618] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc0cd48-32c4-4fc7-99a0-f62f613ea6bf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.199216] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d51e1d-58e7-4e61-b8f5-a24259f9bf45 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.232163] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1a4cb9-3c01-4c93-842a-141399eb2f7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.240041] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522826ed-1321-dcf3-9826-f11852651b6b, 'name': SearchDatastore_Task, 'duration_secs': 0.008826} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.242021] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.242293] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 81f821a3-6238-4ab1-9fe3-56d3f83ad52c/81f821a3-6238-4ab1-9fe3-56d3f83ad52c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.242567] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26a87b7b-b026-4211-8255-d81d10b94845 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.245094] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44590c6c-5a94-43f5-8360-0ba49c487c6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.258780] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.261086] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 996.261086] env[61964]: value = "task-1041429" [ 996.261086] env[61964]: _type = "Task" [ 996.261086] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.269667] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.435373] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.483477] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 996.512573] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 996.512920] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.513485] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 996.513485] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.513763] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 996.513995] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 996.514256] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 996.514434] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 996.514665] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 996.514846] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 996.515066] env[61964]: DEBUG nova.virt.hardware [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 996.515997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9052be76-0816-4912-8057-d2decbd0abaa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.526801] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ef243f-f835-4b9e-9d58-3607560a0c87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.616710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7b04eff1-9178-4000-be72-39fe019f0ac3 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.866s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.639740] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.762533] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 996.774940] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041429, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467629} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.777394] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 81f821a3-6238-4ab1-9fe3-56d3f83ad52c/81f821a3-6238-4ab1-9fe3-56d3f83ad52c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 996.777780] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.780309] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d6b806a-446f-430e-868a-593210a9ff99 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.788619] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 996.788619] env[61964]: value = "task-1041430" [ 996.788619] env[61964]: _type = "Task" [ 996.788619] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.797590] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.819432] env[61964]: DEBUG nova.network.neutron [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating instance_info_cache with network_info: [{"id": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "address": "fa:16:3e:e2:22:9b", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped20338e-f0", "ovs_interfaceid": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.934088] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.072474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.072474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.072474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.072474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.072474] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.075478] env[61964]: INFO nova.compute.manager [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Terminating instance [ 997.272145] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 997.272145] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.316s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.272145] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.996s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.302346] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058483} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.302346] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.303089] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060fad0d-5715-4d7e-9985-ed097257fb65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.331755] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 81f821a3-6238-4ab1-9fe3-56d3f83ad52c/81f821a3-6238-4ab1-9fe3-56d3f83ad52c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.332412] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.333035] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Instance network_info: |[{"id": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "address": "fa:16:3e:e2:22:9b", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped20338e-f0", "ovs_interfaceid": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 997.333711] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bc2cb01-de64-4bea-ba81-b207c2075cd6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.349124] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:22:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbd2870d-a51d-472a-8034-1b3e132b5cb6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed20338e-f092-4780-aede-3f9c39a0b9ea', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.357159] env[61964]: DEBUG oslo.service.loopingcall [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.357488] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 997.360012] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4b4cc17-e195-449b-ab07-3d584caa7135 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.374527] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 997.374527] env[61964]: value = "task-1041431" [ 997.374527] env[61964]: _type = "Task" [ 997.374527] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.379609] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.379609] env[61964]: value = "task-1041432" [ 997.379609] env[61964]: _type = "Task" [ 997.379609] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.383514] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.394514] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041432, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.433606] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.470416] env[61964]: DEBUG nova.compute.manager [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Received event network-changed-ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 997.470416] env[61964]: DEBUG nova.compute.manager [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Refreshing instance network info cache due to event network-changed-ed20338e-f092-4780-aede-3f9c39a0b9ea. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 997.470416] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Acquiring lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.471024] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Acquired lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.471643] env[61964]: DEBUG nova.network.neutron [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Refreshing network info cache for port ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.500984] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Successfully updated port: 453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.579792] env[61964]: DEBUG nova.compute.manager [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 997.579792] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 997.580529] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0aee1b-97a0-41cc-ba15-c87bf1e2baa7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.588520] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 997.588950] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56e912bf-a36c-4f86-9e25-c98ef8a8680d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.594976] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 997.594976] env[61964]: value = "task-1041433" [ 997.594976] env[61964]: _type = "Task" [ 997.594976] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.603339] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041433, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.891318] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041431, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.897338] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041432, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.936454] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.952211] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40df2cf-88ab-40ac-83c3-01867a3d5091 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.960232] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328e3053-e76b-481e-93c7-077753da7b7b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.000446] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2694be26-7065-443c-b49d-21cecd676d1c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.003842] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.003991] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.004178] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.010474] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb42251-9431-4db7-b180-7a660f64ac93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.026345] env[61964]: DEBUG nova.compute.provider_tree [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.104839] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041433, 'name': PowerOffVM_Task, 'duration_secs': 0.437563} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.105380] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.105380] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 998.105572] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e45c521-2c0c-4b58-8b4e-03e5d1393ffa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.181705] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 998.182042] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 998.182366] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Deleting the datastore file [datastore1] 2436ae21-6244-4f8b-abb0-1e7530f07dc5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.182664] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-267fcece-cc1e-4a42-8fb5-41f0aac7960d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.191707] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for the task: (returnval){ [ 998.191707] env[61964]: value = "task-1041435" [ 998.191707] env[61964]: _type = "Task" [ 998.191707] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.199555] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041435, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.221680] env[61964]: DEBUG nova.network.neutron [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updated VIF entry in instance network info cache for port ed20338e-f092-4780-aede-3f9c39a0b9ea. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 998.222058] env[61964]: DEBUG nova.network.neutron [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating instance_info_cache with network_info: [{"id": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "address": "fa:16:3e:e2:22:9b", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped20338e-f0", "ovs_interfaceid": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.386417] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041431, 'name': ReconfigVM_Task, 'duration_secs': 0.772287} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.389751] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 81f821a3-6238-4ab1-9fe3-56d3f83ad52c/81f821a3-6238-4ab1-9fe3-56d3f83ad52c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.390404] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d750a87-ffd6-481b-a692-d47a8c6754a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.396786] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041432, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.397946] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 998.397946] env[61964]: value = "task-1041436" [ 998.397946] env[61964]: _type = "Task" [ 998.397946] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.406252] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041436, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.434067] env[61964]: DEBUG oslo_vmware.api [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041422, 'name': ReconfigVM_Task, 'duration_secs': 5.791633} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.434352] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.434573] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Reconfigured VM to detach interface {{(pid=61964) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 998.530131] env[61964]: DEBUG nova.scheduler.client.report [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 998.570473] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 998.701461] env[61964]: DEBUG oslo_vmware.api [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Task: {'id': task-1041435, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354767} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.701730] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.701922] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 998.702116] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 998.702297] env[61964]: INFO nova.compute.manager [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Took 1.12 seconds to destroy the instance on the hypervisor. [ 998.702542] env[61964]: DEBUG oslo.service.loopingcall [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.702738] env[61964]: DEBUG nova.compute.manager [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 998.702879] env[61964]: DEBUG nova.network.neutron [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 998.725089] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Releasing lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.725319] env[61964]: DEBUG nova.compute.manager [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Received event network-vif-plugged-453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 998.725530] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Acquiring lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.725754] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.725928] env[61964]: DEBUG oslo_concurrency.lockutils [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.726136] env[61964]: DEBUG nova.compute.manager [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] No waiting events found dispatching network-vif-plugged-453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 998.726319] env[61964]: WARNING nova.compute.manager [req-ded609c5-5d16-4b40-b0e4-538a9be1a59c req-87a5a055-1ee8-43d7-a349-efa5755aa7f9 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Received unexpected event network-vif-plugged-453a1893-8117-4c05-9f1c-8ced6b7f897f for instance with vm_state building and task_state spawning. [ 998.794057] env[61964]: DEBUG nova.network.neutron [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Updating instance_info_cache with network_info: [{"id": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "address": "fa:16:3e:2a:04:f3", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap453a1893-81", "ovs_interfaceid": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.896602] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041432, 'name': CreateVM_Task, 'duration_secs': 1.458754} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.900116] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 998.900116] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.900116] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.900116] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.900116] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62c79d11-8234-48a5-af3b-bf1158a0e4a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.909370] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041436, 'name': Rename_Task, 'duration_secs': 0.13514} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.910488] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.910988] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 998.910988] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52117e24-b8fd-598a-0b1e-12ce4539e458" [ 998.910988] env[61964]: _type = "Task" [ 998.910988] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.911446] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c2cb30f-bac2-4de3-a5f6-9d427fd04137 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.927067] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52117e24-b8fd-598a-0b1e-12ce4539e458, 'name': SearchDatastore_Task, 'duration_secs': 0.00888} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.928496] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.928865] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.929197] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.929419] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.929667] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.930059] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 998.930059] env[61964]: value = "task-1041437" [ 998.930059] env[61964]: _type = "Task" [ 998.930059] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.930949] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b5cc4e8-077e-43fe-ab49-62223b72e7c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.944097] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041437, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.945416] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.945495] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 998.947446] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7d6499e-6c27-4232-bb5f-203b99c8f2ac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.953537] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 998.953537] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a63d1-dd9d-2dbb-3b19-0f90952d8f84" [ 998.953537] env[61964]: _type = "Task" [ 998.953537] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.962309] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a63d1-dd9d-2dbb-3b19-0f90952d8f84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.968741] env[61964]: DEBUG nova.compute.manager [req-c66b3384-c6e6-49e8-a4a0-62c429b7ac31 req-b79162bd-b993-462b-84ad-010d44dee9db service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Received event network-vif-deleted-1081d12c-3c4a-471e-b29c-c96b1bfed882 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 998.969110] env[61964]: INFO nova.compute.manager [req-c66b3384-c6e6-49e8-a4a0-62c429b7ac31 req-b79162bd-b993-462b-84ad-010d44dee9db service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Neutron deleted interface 1081d12c-3c4a-471e-b29c-c96b1bfed882; detaching it from the instance and deleting it from the info cache [ 998.969385] env[61964]: DEBUG nova.network.neutron [req-c66b3384-c6e6-49e8-a4a0-62c429b7ac31 req-b79162bd-b993-462b-84ad-010d44dee9db service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.296776] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.297048] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Instance network_info: |[{"id": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "address": "fa:16:3e:2a:04:f3", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap453a1893-81", "ovs_interfaceid": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 999.297528] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:04:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '453a1893-8117-4c05-9f1c-8ced6b7f897f', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.304973] env[61964]: DEBUG oslo.service.loopingcall [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.305220] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 999.305450] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b9ff8c8-aa79-46c7-b23d-9eaabc14db00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.326077] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.326077] env[61964]: value = "task-1041438" [ 999.326077] env[61964]: _type = "Task" [ 999.326077] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.333877] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041438, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.439805] env[61964]: DEBUG nova.network.neutron [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.445012] env[61964]: DEBUG oslo_vmware.api [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041437, 'name': PowerOnVM_Task, 'duration_secs': 0.462091} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.445012] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.445012] env[61964]: INFO nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Took 8.81 seconds to spawn the instance on the hypervisor. [ 999.445216] env[61964]: DEBUG nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 999.445997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7706df-d7af-4bb0-86d3-48f29d8da790 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.462729] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a63d1-dd9d-2dbb-3b19-0f90952d8f84, 'name': SearchDatastore_Task, 'duration_secs': 0.009039} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.463480] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d63089c-202f-4ed9-8874-929cd4f29f71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.468387] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 999.468387] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c0005b-4903-0729-c9c0-2a36248aa346" [ 999.468387] env[61964]: _type = "Task" [ 999.468387] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.471915] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06fe8449-5d5e-4f2f-bd9f-fe9dd4692643 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.478241] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c0005b-4903-0729-c9c0-2a36248aa346, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.481955] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6111d279-41f4-4a27-bd1f-864600ce5618 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.498436] env[61964]: DEBUG nova.compute.manager [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Received event network-changed-453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 999.498646] env[61964]: DEBUG nova.compute.manager [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Refreshing instance network info cache due to event network-changed-453a1893-8117-4c05-9f1c-8ced6b7f897f. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 999.498893] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] Acquiring lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.499052] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] Acquired lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.499739] env[61964]: DEBUG nova.network.neutron [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Refreshing network info cache for port 453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.512917] env[61964]: DEBUG nova.compute.manager [req-c66b3384-c6e6-49e8-a4a0-62c429b7ac31 req-b79162bd-b993-462b-84ad-010d44dee9db service nova] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Detach interface failed, port_id=1081d12c-3c4a-471e-b29c-c96b1bfed882, reason: Instance 2436ae21-6244-4f8b-abb0-1e7530f07dc5 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 999.541659] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.271s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.837393] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041438, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.923086] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.923319] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquired lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.923514] env[61964]: DEBUG nova.network.neutron [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.945819] env[61964]: INFO nova.compute.manager [-] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Took 1.24 seconds to deallocate network for instance. [ 999.961586] env[61964]: INFO nova.compute.manager [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Took 18.23 seconds to build instance. [ 999.979073] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c0005b-4903-0729-c9c0-2a36248aa346, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.979358] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.979618] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 0da26d16-84f2-4c00-97f5-ff132fa122d2/0da26d16-84f2-4c00-97f5-ff132fa122d2.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 999.980379] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-febe6417-27ae-4084-929f-c0f94945cf65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.987455] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 999.987455] env[61964]: value = "task-1041439" [ 999.987455] env[61964]: _type = "Task" [ 999.987455] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.995973] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041439, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.103881] env[61964]: INFO nova.scheduler.client.report [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocation for migration 9a2fc428-7e53-4418-89a7-f123879cb767 [ 1000.261052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.261371] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.261589] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.261776] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.261946] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.264293] env[61964]: INFO nova.compute.manager [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Terminating instance [ 1000.307701] env[61964]: DEBUG nova.network.neutron [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Updated VIF entry in instance network info cache for port 453a1893-8117-4c05-9f1c-8ced6b7f897f. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1000.308126] env[61964]: DEBUG nova.network.neutron [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Updating instance_info_cache with network_info: [{"id": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "address": "fa:16:3e:2a:04:f3", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap453a1893-81", "ovs_interfaceid": "453a1893-8117-4c05-9f1c-8ced6b7f897f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.337421] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041438, 'name': CreateVM_Task, 'duration_secs': 0.711996} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.337612] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1000.338337] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.338513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.338876] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.339174] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b157d47c-ac78-4e4c-9f80-80a8d4f3d95e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.344753] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1000.344753] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41a2d-0624-f10a-500d-61c524a28634" [ 1000.344753] env[61964]: _type = "Task" [ 1000.344753] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.353412] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41a2d-0624-f10a-500d-61c524a28634, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.452821] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.453142] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.453382] env[61964]: DEBUG nova.objects.instance [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lazy-loading 'resources' on Instance uuid 2436ae21-6244-4f8b-abb0-1e7530f07dc5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.463095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ab2c79d8-9dad-40fa-89bd-383803fd1e2a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.745s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.498626] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041439, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437777} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.501085] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 0da26d16-84f2-4c00-97f5-ff132fa122d2/0da26d16-84f2-4c00-97f5-ff132fa122d2.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1000.501648] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.501648] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25b6c33b-e54b-44dd-b93f-542124ee0910 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.507797] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1000.507797] env[61964]: value = "task-1041440" [ 1000.507797] env[61964]: _type = "Task" [ 1000.507797] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.516105] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.516411] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041440, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.609308] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c13a3631-53f7-4495-baeb-052b538ae2ac tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.838s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.613121] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.095s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.613121] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "39683f54-fa99-448d-91ab-171397be48aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.613121] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.613121] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.614424] env[61964]: INFO nova.compute.manager [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Terminating instance [ 1000.721297] env[61964]: INFO nova.network.neutron [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Port e678489f-0505-421c-893c-43523e7e29f7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1000.721676] env[61964]: DEBUG nova.network.neutron [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [{"id": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "address": "fa:16:3e:d1:92:f3", "network": {"id": "48dd64e7-3659-4104-8273-6f58f8d3eb44", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-733934245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d4175848e8e492ba4f749e464b693ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60bdba1a-14cf-46b2-9d8b-aeaf4d80c815", "external-id": "nsx-vlan-transportzone-922", "segmentation_id": 922, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc9cb29f-84", "ovs_interfaceid": "bc9cb29f-849e-4bea-8532-cd8b9efcb198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.768442] env[61964]: DEBUG nova.compute.manager [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1000.769417] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1000.769610] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4f3af2-843f-4e5e-a0c6-4c5ad4847c02 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.777389] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.777620] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03c53fb3-508e-4a0f-89cc-cfd8bdbf8b20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.783949] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 1000.783949] env[61964]: value = "task-1041441" [ 1000.783949] env[61964]: _type = "Task" [ 1000.783949] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.791738] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.811105] env[61964]: DEBUG oslo_concurrency.lockutils [req-e0817c1d-3590-4bbf-af74-4899098f1ce3 req-5414df79-f7a9-4a94-905b-4c0cdc767bef service nova] Releasing lock "refresh_cache-f7f18f7b-5e58-426c-aa5d-7063a50c32ef" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.855972] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d41a2d-0624-f10a-500d-61c524a28634, 'name': SearchDatastore_Task, 'duration_secs': 0.047447} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.856385] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.856628] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.856926] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.857119] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.857340] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.857579] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7a3c5f7-87ae-42c9-acb5-61377ee59eb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.866061] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.866255] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1000.866961] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4ff5b92-7b55-405d-9ef9-ee163b2c019d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.873892] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1000.873892] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234e675-fc3e-d9ba-4b8b-472e6df95e83" [ 1000.873892] env[61964]: _type = "Task" [ 1000.873892] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.880782] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234e675-fc3e-d9ba-4b8b-472e6df95e83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.019105] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041440, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059024} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.019454] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.020294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbc313b-d9e2-48fa-9a21-731c709a1455 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.045103] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 0da26d16-84f2-4c00-97f5-ff132fa122d2/0da26d16-84f2-4c00-97f5-ff132fa122d2.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.047981] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3908102f-4f84-4b53-80a6-3511d2b98197 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.068047] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1001.068047] env[61964]: value = "task-1041442" [ 1001.068047] env[61964]: _type = "Task" [ 1001.068047] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.078274] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041442, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.083670] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.083915] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.119047] env[61964]: DEBUG nova.compute.manager [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1001.119300] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1001.120572] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40749e1-b777-42b2-8671-72e4bd189438 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.130979] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1001.131254] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-342f3404-38ab-4869-b786-4f0154be0c9f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.138073] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1001.138073] env[61964]: value = "task-1041443" [ 1001.138073] env[61964]: _type = "Task" [ 1001.138073] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.146211] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.148023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e7631f-9e17-4661-b1a1-adca3498be6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.155010] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfba0514-2510-4f1c-8acb-b6beaf6bef0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.187408] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ad3c4e-5bd9-465a-ac65-18f32c5a4675 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.195906] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6402e5-1af6-48b1-80d9-e2342e98e008 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.211535] env[61964]: DEBUG nova.compute.provider_tree [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.224195] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Releasing lock "refresh_cache-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.294841] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041441, 'name': PowerOffVM_Task, 'duration_secs': 0.201407} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.295163] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1001.295892] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1001.295892] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78e03043-c908-4250-bc8a-1d67a9ffe3ea {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.384573] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5234e675-fc3e-d9ba-4b8b-472e6df95e83, 'name': SearchDatastore_Task, 'duration_secs': 0.009614} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.385445] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f57ff1bb-188b-454c-b6c1-fd74a1add51d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.391167] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1001.391167] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288f6e2-3f40-3f34-9a70-f07552aea30d" [ 1001.391167] env[61964]: _type = "Task" [ 1001.391167] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.399346] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288f6e2-3f40-3f34-9a70-f07552aea30d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.411883] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.412169] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.412364] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleting the datastore file [datastore2] c4fabe85-1b95-4981-9e05-dbfedbbe1e1e {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.412683] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-033a0931-b687-40d9-b334-ae315088bdc9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.419206] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.419437] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.419646] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.419840] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.420031] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.423090] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 1001.423090] env[61964]: value = "task-1041445" [ 1001.423090] env[61964]: _type = "Task" [ 1001.423090] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.423574] env[61964]: INFO nova.compute.manager [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Terminating instance [ 1001.433872] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.577989] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041442, 'name': ReconfigVM_Task, 'duration_secs': 0.37353} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.578137] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 0da26d16-84f2-4c00-97f5-ff132fa122d2/0da26d16-84f2-4c00-97f5-ff132fa122d2.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.578822] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8fe4e34-b464-45a5-a42c-33f767f1d741 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.584651] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1001.584651] env[61964]: value = "task-1041446" [ 1001.584651] env[61964]: _type = "Task" [ 1001.584651] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.587861] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1001.595259] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041446, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.648016] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041443, 'name': PowerOffVM_Task, 'duration_secs': 0.161042} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.648365] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1001.648426] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1001.648680] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92ed7126-152b-43aa-9b38-0cd78e5c58eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.714257] env[61964]: DEBUG nova.scheduler.client.report [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1001.727962] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8f24f825-d221-44a6-b083-030a959fe36e tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "interface-c4fabe85-1b95-4981-9e05-dbfedbbe1e1e-e678489f-0505-421c-893c-43523e7e29f7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.877s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.767563] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.767819] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.768138] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] 39683f54-fa99-448d-91ab-171397be48aa {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.768388] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee78771d-54c5-4124-91ea-e802d236f56a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.774795] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1001.774795] env[61964]: value = "task-1041448" [ 1001.774795] env[61964]: _type = "Task" [ 1001.774795] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.782415] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.902096] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5288f6e2-3f40-3f34-9a70-f07552aea30d, 'name': SearchDatastore_Task, 'duration_secs': 0.018119} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.902356] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.902638] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f7f18f7b-5e58-426c-aa5d-7063a50c32ef/f7f18f7b-5e58-426c-aa5d-7063a50c32ef.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1001.902927] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ed808de-1937-4b98-8387-eefd6fb909aa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.909556] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1001.909556] env[61964]: value = "task-1041449" [ 1001.909556] env[61964]: _type = "Task" [ 1001.909556] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.919020] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.933387] env[61964]: DEBUG nova.compute.manager [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1001.933668] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1001.933962] env[61964]: DEBUG oslo_vmware.api [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208753} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.934673] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48454ead-113f-4828-aa08-38a258df4c7f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.937492] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.937681] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1001.937900] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1001.938092] env[61964]: INFO nova.compute.manager [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1001.938355] env[61964]: DEBUG oslo.service.loopingcall [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.938557] env[61964]: DEBUG nova.compute.manager [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1001.938642] env[61964]: DEBUG nova.network.neutron [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1001.944271] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1001.944520] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74b8a199-7ca7-40cc-b5d2-e6401a11aa20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.955120] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1001.955120] env[61964]: value = "task-1041450" [ 1001.955120] env[61964]: _type = "Task" [ 1001.955120] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.966258] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.097325] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041446, 'name': Rename_Task, 'duration_secs': 0.150553} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.100544] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.101689] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0421235c-2db7-464b-be06-62d63ef5bb10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.108208] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1002.108208] env[61964]: value = "task-1041451" [ 1002.108208] env[61964]: _type = "Task" [ 1002.108208] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.117149] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.125092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.219461] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.222097] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.100s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.223635] env[61964]: INFO nova.compute.claims [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.244668] env[61964]: INFO nova.scheduler.client.report [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Deleted allocations for instance 2436ae21-6244-4f8b-abb0-1e7530f07dc5 [ 1002.286809] env[61964]: DEBUG oslo_vmware.api [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141577} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.287203] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.287347] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1002.287534] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1002.287856] env[61964]: INFO nova.compute.manager [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1002.287998] env[61964]: DEBUG oslo.service.loopingcall [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.288458] env[61964]: DEBUG nova.compute.manager [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1002.288458] env[61964]: DEBUG nova.network.neutron [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1002.419101] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492269} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.419497] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] f7f18f7b-5e58-426c-aa5d-7063a50c32ef/f7f18f7b-5e58-426c-aa5d-7063a50c32ef.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1002.419599] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.419851] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5e3ed0b-f5d0-4b8a-9883-d71de1eb65cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.425902] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1002.425902] env[61964]: value = "task-1041452" [ 1002.425902] env[61964]: _type = "Task" [ 1002.425902] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.433742] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041452, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.465462] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041450, 'name': PowerOffVM_Task, 'duration_secs': 0.23007} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.465803] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1002.466020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1002.466338] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20697808-5f36-4331-b596-b88c3a9db8f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.541119] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1002.542331] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1002.542738] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore1] 81f821a3-6238-4ab1-9fe3-56d3f83ad52c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.542983] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cb9a3b6-40e5-4c48-8d14-912218ebc1a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.550802] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1002.550802] env[61964]: value = "task-1041454" [ 1002.550802] env[61964]: _type = "Task" [ 1002.550802] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.559320] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.617947] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041451, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.631413] env[61964]: DEBUG nova.compute.manager [req-fe1d2a0e-6eca-4ed8-9785-1bb4a883d1ed req-8bdf88ad-bc34-48bc-85e7-595691e795e1 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Received event network-vif-deleted-79400d4a-e480-4d60-8c9f-dcb1b899ae7c {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1002.631664] env[61964]: INFO nova.compute.manager [req-fe1d2a0e-6eca-4ed8-9785-1bb4a883d1ed req-8bdf88ad-bc34-48bc-85e7-595691e795e1 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Neutron deleted interface 79400d4a-e480-4d60-8c9f-dcb1b899ae7c; detaching it from the instance and deleting it from the info cache [ 1002.631922] env[61964]: DEBUG nova.network.neutron [req-fe1d2a0e-6eca-4ed8-9785-1bb4a883d1ed req-8bdf88ad-bc34-48bc-85e7-595691e795e1 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.744613] env[61964]: DEBUG nova.compute.manager [req-9aa41479-4674-409a-96c0-03f9d7b6f78e req-d94b2132-b51e-4f50-bf2e-cd614df57761 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Received event network-vif-deleted-bc9cb29f-849e-4bea-8532-cd8b9efcb198 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1002.744905] env[61964]: INFO nova.compute.manager [req-9aa41479-4674-409a-96c0-03f9d7b6f78e req-d94b2132-b51e-4f50-bf2e-cd614df57761 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Neutron deleted interface bc9cb29f-849e-4bea-8532-cd8b9efcb198; detaching it from the instance and deleting it from the info cache [ 1002.745083] env[61964]: DEBUG nova.network.neutron [req-9aa41479-4674-409a-96c0-03f9d7b6f78e req-d94b2132-b51e-4f50-bf2e-cd614df57761 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.755909] env[61964]: DEBUG oslo_concurrency.lockutils [None req-95272e04-cae0-4f8a-9759-b76f6c5033c2 tempest-ImagesNegativeTestJSON-1743430746 tempest-ImagesNegativeTestJSON-1743430746-project-member] Lock "2436ae21-6244-4f8b-abb0-1e7530f07dc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.684s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.936145] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041452, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080645} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.936594] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1002.937518] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2205466e-b089-484d-baa5-27c97074d343 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.964015] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] f7f18f7b-5e58-426c-aa5d-7063a50c32ef/f7f18f7b-5e58-426c-aa5d-7063a50c32ef.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.964015] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c768caa5-6257-4fa3-9da4-b9ac7f95f1d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.986020] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1002.986020] env[61964]: value = "task-1041455" [ 1002.986020] env[61964]: _type = "Task" [ 1002.986020] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.992389] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041455, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.060468] env[61964]: DEBUG oslo_vmware.api [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238104} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.060744] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.060940] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1003.061184] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1003.061380] env[61964]: INFO nova.compute.manager [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1003.061630] env[61964]: DEBUG oslo.service.loopingcall [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.061829] env[61964]: DEBUG nova.compute.manager [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1003.061931] env[61964]: DEBUG nova.network.neutron [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1003.109398] env[61964]: DEBUG nova.network.neutron [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.123640] env[61964]: DEBUG oslo_vmware.api [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041451, 'name': PowerOnVM_Task, 'duration_secs': 0.678893} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.123932] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1003.124180] env[61964]: INFO nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Took 8.42 seconds to spawn the instance on the hypervisor. [ 1003.124368] env[61964]: DEBUG nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1003.125246] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5ab130-082b-493d-b29f-8cdd672dabad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.135533] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66b92492-d541-438d-a2d4-1cb6197b7879 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.144851] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7949d3c5-0714-42f7-862e-37baa85517bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.177530] env[61964]: DEBUG nova.compute.manager [req-fe1d2a0e-6eca-4ed8-9785-1bb4a883d1ed req-8bdf88ad-bc34-48bc-85e7-595691e795e1 service nova] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Detach interface failed, port_id=79400d4a-e480-4d60-8c9f-dcb1b899ae7c, reason: Instance 39683f54-fa99-448d-91ab-171397be48aa could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1003.221968] env[61964]: DEBUG nova.network.neutron [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.248869] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05104f3d-a971-4eb0-85f5-f4f5200da7e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.263273] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9128b98d-7171-4904-980d-b32444ac9187 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.290876] env[61964]: DEBUG nova.compute.manager [req-9aa41479-4674-409a-96c0-03f9d7b6f78e req-d94b2132-b51e-4f50-bf2e-cd614df57761 service nova] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Detach interface failed, port_id=bc9cb29f-849e-4bea-8532-cd8b9efcb198, reason: Instance c4fabe85-1b95-4981-9e05-dbfedbbe1e1e could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1003.429372] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768a36bf-88c5-4acb-8098-d07a8fb46726 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.437127] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aef1b9c-6c59-4c75-beed-d12db6b18cab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.470213] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfe5e2e-2644-49ec-b96b-fa3d049c74b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.477615] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01404161-5fa0-4267-8312-38ad852b15d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.492575] env[61964]: DEBUG nova.compute.provider_tree [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1003.501530] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041455, 'name': ReconfigVM_Task, 'duration_secs': 0.267758} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.501922] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Reconfigured VM instance instance-00000062 to attach disk [datastore2] f7f18f7b-5e58-426c-aa5d-7063a50c32ef/f7f18f7b-5e58-426c-aa5d-7063a50c32ef.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.502456] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-617d0cc0-115b-4bbe-91bd-1b07fe8186c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.508360] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1003.508360] env[61964]: value = "task-1041456" [ 1003.508360] env[61964]: _type = "Task" [ 1003.508360] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.521198] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041456, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.615497] env[61964]: INFO nova.compute.manager [-] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Took 1.33 seconds to deallocate network for instance. [ 1003.644377] env[61964]: INFO nova.compute.manager [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Took 17.85 seconds to build instance. [ 1003.724188] env[61964]: INFO nova.compute.manager [-] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Took 1.79 seconds to deallocate network for instance. [ 1003.875032] env[61964]: DEBUG nova.network.neutron [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.018922] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041456, 'name': Rename_Task, 'duration_secs': 0.160287} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.020022] env[61964]: ERROR nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [req-b65ae3ce-8e52-4635-91ff-01409276290d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b65ae3ce-8e52-4635-91ff-01409276290d"}]} [ 1004.020497] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1004.022945] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2edf927c-b5ea-4892-9dee-50e37bbc0adf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.031081] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1004.031081] env[61964]: value = "task-1041457" [ 1004.031081] env[61964]: _type = "Task" [ 1004.031081] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.039534] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.040842] env[61964]: DEBUG nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1004.057528] env[61964]: DEBUG nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1004.057756] env[61964]: DEBUG nova.compute.provider_tree [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.071627] env[61964]: DEBUG nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1004.094606] env[61964]: DEBUG nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1004.124723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.149800] env[61964]: DEBUG oslo_concurrency.lockutils [None req-592399e0-4b9f-41ee-ba3b-f784b9e2390a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.389s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.234067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.277172] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5e6335-a5c4-4d17-be32-664ef84d2421 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.285260] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12b0c99-8685-412c-ab3b-51f931df5ff5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.318902] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d84efa-c621-4c7b-b98a-069e7898c18b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.326889] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8740297e-177e-4fa7-8449-877acf8d1554 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.340886] env[61964]: DEBUG nova.compute.provider_tree [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.377259] env[61964]: INFO nova.compute.manager [-] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Took 1.32 seconds to deallocate network for instance. [ 1004.543226] env[61964]: DEBUG oslo_vmware.api [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041457, 'name': PowerOnVM_Task, 'duration_secs': 0.469118} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.543514] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1004.543718] env[61964]: INFO nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1004.543987] env[61964]: DEBUG nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1004.544844] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683f39a2-8cf8-4279-be82-d201b9216b4c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.777840] env[61964]: DEBUG nova.compute.manager [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Received event network-vif-deleted-4d6064e2-9257-47a1-93d4-5208c08288f2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1004.777951] env[61964]: DEBUG nova.compute.manager [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Received event network-changed-ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1004.778128] env[61964]: DEBUG nova.compute.manager [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Refreshing instance network info cache due to event network-changed-ed20338e-f092-4780-aede-3f9c39a0b9ea. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1004.778360] env[61964]: DEBUG oslo_concurrency.lockutils [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] Acquiring lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.778517] env[61964]: DEBUG oslo_concurrency.lockutils [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] Acquired lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.778684] env[61964]: DEBUG nova.network.neutron [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Refreshing network info cache for port ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.875023] env[61964]: DEBUG nova.scheduler.client.report [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 128 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1004.875023] env[61964]: DEBUG nova.compute.provider_tree [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 128 to 129 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1004.875023] env[61964]: DEBUG nova.compute.provider_tree [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.884573] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.061778] env[61964]: INFO nova.compute.manager [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Took 18.60 seconds to build instance. [ 1005.378964] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.157s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.379634] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1005.382663] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.258s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.382873] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.386913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.152s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.386913] env[61964]: DEBUG nova.objects.instance [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'resources' on Instance uuid c4fabe85-1b95-4981-9e05-dbfedbbe1e1e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.410436] env[61964]: INFO nova.scheduler.client.report [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocations for instance 39683f54-fa99-448d-91ab-171397be48aa [ 1005.563411] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7eaec247-e8dc-446c-926e-fecc046e6c9c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.106s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.612199] env[61964]: DEBUG nova.network.neutron [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updated VIF entry in instance network info cache for port ed20338e-f092-4780-aede-3f9c39a0b9ea. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1005.612585] env[61964]: DEBUG nova.network.neutron [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating instance_info_cache with network_info: [{"id": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "address": "fa:16:3e:e2:22:9b", "network": {"id": "68c8eecb-0e06-43ac-b970-9e93196d39a3", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1153431131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6110872ae3dc4491bb10abb1945ffe2a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbd2870d-a51d-472a-8034-1b3e132b5cb6", "external-id": "nsx-vlan-transportzone-101", "segmentation_id": 101, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped20338e-f0", "ovs_interfaceid": "ed20338e-f092-4780-aede-3f9c39a0b9ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.851380] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c91499-1bcb-45f2-b110-22dab9508c14 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.856692] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Suspending the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1005.857147] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d6c72136-723b-476b-b657-653010e7bae8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.864299] env[61964]: DEBUG oslo_vmware.api [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1005.864299] env[61964]: value = "task-1041458" [ 1005.864299] env[61964]: _type = "Task" [ 1005.864299] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.876810] env[61964]: DEBUG oslo_vmware.api [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041458, 'name': SuspendVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.889072] env[61964]: DEBUG nova.compute.utils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.895020] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1005.895020] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.925103] env[61964]: DEBUG oslo_concurrency.lockutils [None req-22bc30c4-d5f6-4491-a746-1cdbfefef423 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "39683f54-fa99-448d-91ab-171397be48aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.314s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.936334] env[61964]: DEBUG nova.policy [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f438e739578a4fabad898ebd5e2925cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ccf1fe5003e484fa73ce5c32767479f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.039585] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace4549d-5d1b-4e6e-9971-0a0c0600ccad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.047621] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce7f7db-887e-4fa2-8c08-b6122cd04f6e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.077591] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93403dd-eb76-426d-a0a2-902c14be8f1e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.085771] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe0adfe-91a7-42cd-97fd-4ad1263c328a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.100208] env[61964]: DEBUG nova.compute.provider_tree [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.115064] env[61964]: DEBUG oslo_concurrency.lockutils [req-6176da59-91fc-435f-8096-3d42d8f8333d req-f508fe44-d5d0-4811-892e-e31f7efd5764 service nova] Releasing lock "refresh_cache-0da26d16-84f2-4c00-97f5-ff132fa122d2" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.374489] env[61964]: DEBUG oslo_vmware.api [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041458, 'name': SuspendVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.394990] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1006.438903] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Successfully created port: e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.603886] env[61964]: DEBUG nova.scheduler.client.report [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.876040] env[61964]: DEBUG oslo_vmware.api [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041458, 'name': SuspendVM_Task, 'duration_secs': 0.62871} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.876175] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Suspended the VM {{(pid=61964) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1006.876394] env[61964]: DEBUG nova.compute.manager [None req-3e0331f1-2c79-41b7-9f7b-c96cb708eeed tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.877187] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3aad463-53a8-4024-8a17-bd67d6af6d4e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.112219] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.724s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.113436] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.229s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.113838] env[61964]: DEBUG nova.objects.instance [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid 81f821a3-6238-4ab1-9fe3-56d3f83ad52c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.135908] env[61964]: INFO nova.scheduler.client.report [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted allocations for instance c4fabe85-1b95-4981-9e05-dbfedbbe1e1e [ 1007.405437] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1007.433155] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1007.434020] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.434020] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1007.434020] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.434020] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1007.434020] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1007.434240] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1007.434406] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1007.434625] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1007.434750] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1007.434941] env[61964]: DEBUG nova.virt.hardware [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1007.436047] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe2e719-82d1-410c-8745-6c01fcdf94f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.444289] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f966c7a9-4428-4e6e-a25f-807b6b998046 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.646723] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d1a9dde6-4ae4-4896-a4a5-e21b402ee300 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "c4fabe85-1b95-4981-9e05-dbfedbbe1e1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.385s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.772977] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1cad3f-cf66-40b7-9b15-fc4e35287e04 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.783489] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695c35dd-7cca-40db-b964-559d2b63e40f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.819493] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01912c68-ebf4-47e8-bc88-cfd46d1aa247 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.833625] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11462b5-842e-4a64-a205-a6ad14397e3e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.850372] env[61964]: DEBUG nova.compute.provider_tree [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.981406] env[61964]: DEBUG nova.compute.manager [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Received event network-vif-plugged-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1007.981670] env[61964]: DEBUG oslo_concurrency.lockutils [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] Acquiring lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.981848] env[61964]: DEBUG oslo_concurrency.lockutils [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.982015] env[61964]: DEBUG oslo_concurrency.lockutils [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.982257] env[61964]: DEBUG nova.compute.manager [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] No waiting events found dispatching network-vif-plugged-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1007.982418] env[61964]: WARNING nova.compute.manager [req-da706e03-109f-4720-b698-c49f6e1b1161 req-0ae4829f-26af-431f-8457-0d8a850a1abe service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Received unexpected event network-vif-plugged-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 for instance with vm_state building and task_state spawning. [ 1008.052116] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.052479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.052706] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.052900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.053095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.055706] env[61964]: INFO nova.compute.manager [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Terminating instance [ 1008.079704] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Successfully updated port: e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.354115] env[61964]: DEBUG nova.scheduler.client.report [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1008.425925] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.426254] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.426480] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.426710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.426843] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.431676] env[61964]: INFO nova.compute.manager [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Terminating instance [ 1008.559880] env[61964]: DEBUG nova.compute.manager [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1008.560130] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1008.561080] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cd5432-775a-42d7-aac3-6528f8791060 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.569303] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1008.569539] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5a84d8f-a139-4c22-8ef3-8793f5c1a02b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.576624] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 1008.576624] env[61964]: value = "task-1041459" [ 1008.576624] env[61964]: _type = "Task" [ 1008.576624] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.584083] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.584221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.584397] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.585495] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.859337] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.746s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.885790] env[61964]: INFO nova.scheduler.client.report [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance 81f821a3-6238-4ab1-9fe3-56d3f83ad52c [ 1008.937330] env[61964]: DEBUG nova.compute.manager [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1008.937330] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1008.938050] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc066de-7866-47f2-a869-b83f84d0d002 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.947428] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1008.947707] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4de2a4e-f957-4c87-adfd-1d8a7082fde2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.011956] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1009.012389] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1009.012459] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore2] f7f18f7b-5e58-426c-aa5d-7063a50c32ef {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.012996] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-505054b8-856d-40a6-a6b8-c6c88566ffcb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.019396] env[61964]: DEBUG oslo_vmware.api [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1009.019396] env[61964]: value = "task-1041461" [ 1009.019396] env[61964]: _type = "Task" [ 1009.019396] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.026646] env[61964]: DEBUG oslo_vmware.api [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.085941] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041459, 'name': PowerOffVM_Task, 'duration_secs': 0.182048} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.088120] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1009.088392] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1009.088662] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7546982-081c-4fad-a6c7-7724ede492b5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.117450] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1009.148092] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1009.148390] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1009.148592] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleting the datastore file [datastore2] 47378856-4cbf-4cf6-aecd-ae935885df3d {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.148862] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ba908f9-3cd3-4062-aab7-1d4ede9e6eec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.154773] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for the task: (returnval){ [ 1009.154773] env[61964]: value = "task-1041463" [ 1009.154773] env[61964]: _type = "Task" [ 1009.154773] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.166625] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.292928] env[61964]: DEBUG nova.network.neutron [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Updating instance_info_cache with network_info: [{"id": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "address": "fa:16:3e:5c:ac:cf", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6c0d0c2-f3", "ovs_interfaceid": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.369040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.369040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.394579] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c49d752a-63e8-4182-8802-2df22061b94b tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "81f821a3-6238-4ab1-9fe3-56d3f83ad52c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.975s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.529040] env[61964]: DEBUG oslo_vmware.api [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149891} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.529340] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.529533] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1009.529715] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1009.529893] env[61964]: INFO nova.compute.manager [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1009.530152] env[61964]: DEBUG oslo.service.loopingcall [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.530439] env[61964]: DEBUG nova.compute.manager [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1009.530538] env[61964]: DEBUG nova.network.neutron [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1009.665725] env[61964]: DEBUG oslo_vmware.api [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Task: {'id': task-1041463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10392} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.666012] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.666216] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1009.666434] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1009.666679] env[61964]: INFO nova.compute.manager [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1009.666938] env[61964]: DEBUG oslo.service.loopingcall [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.667149] env[61964]: DEBUG nova.compute.manager [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1009.667247] env[61964]: DEBUG nova.network.neutron [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1009.797584] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.797965] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Instance network_info: |[{"id": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "address": "fa:16:3e:5c:ac:cf", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6c0d0c2-f3", "ovs_interfaceid": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1009.798468] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:ac:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.811138] env[61964]: DEBUG oslo.service.loopingcall [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.811706] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1009.811954] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0946b04a-76ee-4974-b38a-e79dd836aefc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.832265] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.832265] env[61964]: value = "task-1041464" [ 1009.832265] env[61964]: _type = "Task" [ 1009.832265] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.840071] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041464, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.871256] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1009.982521] env[61964]: DEBUG nova.compute.manager [req-0a57825b-2bf4-410b-a89a-19f051051ed7 req-97187f77-5948-4b39-9968-862b7ed3cb90 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Received event network-vif-deleted-083961bb-ae89-4d60-8ea8-3d5454613407 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1009.982716] env[61964]: INFO nova.compute.manager [req-0a57825b-2bf4-410b-a89a-19f051051ed7 req-97187f77-5948-4b39-9968-862b7ed3cb90 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Neutron deleted interface 083961bb-ae89-4d60-8ea8-3d5454613407; detaching it from the instance and deleting it from the info cache [ 1009.983283] env[61964]: DEBUG nova.network.neutron [req-0a57825b-2bf4-410b-a89a-19f051051ed7 req-97187f77-5948-4b39-9968-862b7ed3cb90 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.016942] env[61964]: DEBUG nova.compute.manager [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Received event network-changed-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1010.017197] env[61964]: DEBUG nova.compute.manager [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Refreshing instance network info cache due to event network-changed-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1010.018625] env[61964]: DEBUG oslo_concurrency.lockutils [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] Acquiring lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.018806] env[61964]: DEBUG oslo_concurrency.lockutils [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] Acquired lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.018984] env[61964]: DEBUG nova.network.neutron [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Refreshing network info cache for port e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.266216] env[61964]: DEBUG nova.network.neutron [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.343074] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041464, 'name': CreateVM_Task, 'duration_secs': 0.329002} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.343074] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.343378] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.343666] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.344462] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.345421] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7754e68-5a3b-4f96-8cc0-5f13e6d2cbf5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.350046] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1010.350046] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529960e5-6c4f-42cd-98e7-ac78a5385264" [ 1010.350046] env[61964]: _type = "Task" [ 1010.350046] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.358374] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529960e5-6c4f-42cd-98e7-ac78a5385264, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.390115] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.390375] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.392070] env[61964]: INFO nova.compute.claims [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.413764] env[61964]: DEBUG nova.network.neutron [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.485572] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fdbd893-384b-4f4b-9519-a2f55b4a9a28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.497589] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e284246b-0d85-4317-a464-8233d44245c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.529967] env[61964]: DEBUG nova.compute.manager [req-0a57825b-2bf4-410b-a89a-19f051051ed7 req-97187f77-5948-4b39-9968-862b7ed3cb90 service nova] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Detach interface failed, port_id=083961bb-ae89-4d60-8ea8-3d5454613407, reason: Instance 47378856-4cbf-4cf6-aecd-ae935885df3d could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1010.736232] env[61964]: DEBUG nova.network.neutron [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Updated VIF entry in instance network info cache for port e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.736850] env[61964]: DEBUG nova.network.neutron [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Updating instance_info_cache with network_info: [{"id": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "address": "fa:16:3e:5c:ac:cf", "network": {"id": "9280fea4-ff10-4e29-82cd-119406c4ff05", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-793688596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ccf1fe5003e484fa73ce5c32767479f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6c0d0c2-f3", "ovs_interfaceid": "e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.768056] env[61964]: INFO nova.compute.manager [-] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Took 1.24 seconds to deallocate network for instance. [ 1010.819922] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "3dc30299-8462-4714-8e1c-8a10412e84d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.820179] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.860617] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]529960e5-6c4f-42cd-98e7-ac78a5385264, 'name': SearchDatastore_Task, 'duration_secs': 0.009645} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.860855] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.861109] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.861355] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.861509] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.861693] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.861952] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-900c7ede-a4b5-48a5-9937-5e99a67af067 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.877019] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.877217] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.877934] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1810d774-6489-47a8-abdb-455b8f7e7655 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.883662] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1010.883662] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229e6f2-ae85-8524-aea4-1e90cb34b875" [ 1010.883662] env[61964]: _type = "Task" [ 1010.883662] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.891360] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229e6f2-ae85-8524-aea4-1e90cb34b875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.916597] env[61964]: INFO nova.compute.manager [-] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Took 1.25 seconds to deallocate network for instance. [ 1011.240623] env[61964]: DEBUG oslo_concurrency.lockutils [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] Releasing lock "refresh_cache-0c1e5fc1-4e25-489f-b640-253ce03139d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.240960] env[61964]: DEBUG nova.compute.manager [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Received event network-vif-deleted-453a1893-8117-4c05-9f1c-8ced6b7f897f {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1011.241081] env[61964]: INFO nova.compute.manager [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Neutron deleted interface 453a1893-8117-4c05-9f1c-8ced6b7f897f; detaching it from the instance and deleting it from the info cache [ 1011.241261] env[61964]: DEBUG nova.network.neutron [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.274452] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.323441] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1011.394805] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5229e6f2-ae85-8524-aea4-1e90cb34b875, 'name': SearchDatastore_Task, 'duration_secs': 0.02304} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.395665] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee1a733f-a2f6-475f-ac1c-d005ed79c74b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.403347] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1011.403347] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526eabe3-46e2-4bb6-ad24-dc5c7e43c804" [ 1011.403347] env[61964]: _type = "Task" [ 1011.403347] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.412064] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526eabe3-46e2-4bb6-ad24-dc5c7e43c804, 'name': SearchDatastore_Task, 'duration_secs': 0.008457} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.412294] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.412580] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0c1e5fc1-4e25-489f-b640-253ce03139d5/0c1e5fc1-4e25-489f-b640-253ce03139d5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.412870] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-086c336c-3d12-4ed7-9acc-f67a0ce1153e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.419418] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1011.419418] env[61964]: value = "task-1041465" [ 1011.419418] env[61964]: _type = "Task" [ 1011.419418] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.422514] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.430475] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.533030] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d010a965-af12-4913-862a-3dc491b7427a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.541060] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f83b24-f4ff-4c57-b3b8-cc320d8cc6bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.571306] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dab3e0c-1b25-402e-81c6-5b1f55eed811 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.580978] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6ccebe-a561-4ba1-97d2-6a7e72f59dac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.598909] env[61964]: DEBUG nova.compute.provider_tree [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.743727] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4c98489-6ba2-490b-aa01-2982d3d6bfb8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.754276] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0c27dc-9c5e-4f54-93c1-eb6aa93237f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.784987] env[61964]: DEBUG nova.compute.manager [req-ee386eb2-bbcc-42d3-8656-9d65ed2e7724 req-f09f9c99-4c92-4289-977f-db49d6a81984 service nova] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Detach interface failed, port_id=453a1893-8117-4c05-9f1c-8ced6b7f897f, reason: Instance f7f18f7b-5e58-426c-aa5d-7063a50c32ef could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1011.842754] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.854129] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.854347] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.931326] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041465, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.102521] env[61964]: DEBUG nova.scheduler.client.report [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.357581] env[61964]: DEBUG nova.compute.utils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.432250] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558691} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.432526] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 0c1e5fc1-4e25-489f-b640-253ce03139d5/0c1e5fc1-4e25-489f-b640-253ce03139d5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.432737] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.432987] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfa8d969-5382-4211-9e4a-5b7cd9a81b9b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.439657] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1012.439657] env[61964]: value = "task-1041466" [ 1012.439657] env[61964]: _type = "Task" [ 1012.439657] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.447437] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041466, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.606821] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.607376] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1012.610244] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.336s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.610474] env[61964]: DEBUG nova.objects.instance [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'resources' on Instance uuid f7f18f7b-5e58-426c-aa5d-7063a50c32ef {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.861603] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.950337] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041466, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061043} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.950807] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.951685] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48103607-a3c6-4dfd-bf8e-71dbbad746f8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.974105] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 0c1e5fc1-4e25-489f-b640-253ce03139d5/0c1e5fc1-4e25-489f-b640-253ce03139d5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.974461] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ee5e495-4340-4b84-b66a-5a41df1b39c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.994278] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1012.994278] env[61964]: value = "task-1041467" [ 1012.994278] env[61964]: _type = "Task" [ 1012.994278] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.002796] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041467, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.114620] env[61964]: DEBUG nova.compute.utils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.120607] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1013.120911] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1013.173559] env[61964]: DEBUG nova.policy [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d6a5aa43ac34579b8cee43cc6ac9e2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d63f8afe09843efb54de6ea85f276aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.278991] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ba59db-541d-495a-9568-f05522ce9acb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.287221] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdc891d-4b3f-4579-be49-6bbcf4b014a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.317930] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc79d0de-88c5-454b-8436-b0ca6301cded {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.325791] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc920cf3-181e-4ae3-afa6-84ae56118fec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.339714] env[61964]: DEBUG nova.compute.provider_tree [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1013.490685] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Successfully created port: d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.504490] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041467, 'name': ReconfigVM_Task, 'duration_secs': 0.281722} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.504777] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 0c1e5fc1-4e25-489f-b640-253ce03139d5/0c1e5fc1-4e25-489f-b640-253ce03139d5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.505575] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b41d05d-b1da-4577-befd-9147d86a7b64 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.511588] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1013.511588] env[61964]: value = "task-1041468" [ 1013.511588] env[61964]: _type = "Task" [ 1013.511588] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.520623] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041468, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.621712] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1013.859605] env[61964]: ERROR nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [req-193a0064-e559-4ece-8bd2-20122a7713e8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-193a0064-e559-4ece-8bd2-20122a7713e8"}]} [ 1013.877411] env[61964]: DEBUG nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1013.892573] env[61964]: DEBUG nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1013.892875] env[61964]: DEBUG nova.compute.provider_tree [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1013.904730] env[61964]: DEBUG nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1013.924735] env[61964]: DEBUG nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1013.938070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.938070] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.938299] env[61964]: INFO nova.compute.manager [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Attaching volume b92af457-aeb9-4914-a6b6-8291bbfc0491 to /dev/sdb [ 1013.972821] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a5b4e9-d9e0-46ac-b834-4dad3c7a8ebb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.983336] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585e4ef0-6ed3-429e-b04a-1f7d7cd85e1e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.997686] env[61964]: DEBUG nova.virt.block_device [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating existing volume attachment record: 4c030cf8-b2d3-40fc-a954-a2beff6f9d48 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1014.023810] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041468, 'name': Rename_Task, 'duration_secs': 0.146332} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.024123] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.024390] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb8dd549-9d1c-474a-be35-0f664a80660e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.030924] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1014.030924] env[61964]: value = "task-1041469" [ 1014.030924] env[61964]: _type = "Task" [ 1014.030924] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.039361] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.067257] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95af185-8da3-4d56-b0cd-4e94666cc3b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.075086] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ba6359-8339-4398-b31a-97565f01ab98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.106385] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30d8f0b-bf28-44d1-98f7-1275799c1935 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.113682] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b275b1-9b23-4c58-8a8e-e2c718c9e79d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.131152] env[61964]: DEBUG nova.compute.provider_tree [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.543097] env[61964]: DEBUG oslo_vmware.api [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041469, 'name': PowerOnVM_Task, 'duration_secs': 0.426743} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.543657] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1014.543657] env[61964]: INFO nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1014.544073] env[61964]: DEBUG nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1014.544572] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea20135b-463f-4b8d-b5b4-1d8135dc8d8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.634314] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1014.663202] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1014.663496] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.663719] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1014.663953] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.664188] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1014.664398] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1014.664591] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1014.664759] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1014.664940] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1014.665181] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1014.665365] env[61964]: DEBUG nova.virt.hardware [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1014.666270] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e8439f-eb54-4368-ad93-f84d49ebc973 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.669603] env[61964]: DEBUG nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1014.669825] env[61964]: DEBUG nova.compute.provider_tree [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 131 to 132 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1014.670032] env[61964]: DEBUG nova.compute.provider_tree [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.678511] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d22fd1-4e41-4123-83b2-66bdf0f192fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.905901] env[61964]: DEBUG nova.compute.manager [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1014.906168] env[61964]: DEBUG oslo_concurrency.lockutils [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.906383] env[61964]: DEBUG oslo_concurrency.lockutils [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.906689] env[61964]: DEBUG oslo_concurrency.lockutils [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.906820] env[61964]: DEBUG nova.compute.manager [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] No waiting events found dispatching network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1014.906972] env[61964]: WARNING nova.compute.manager [req-a46a21d4-d1a6-44cf-8088-2bc64aabfb94 req-c5cbd2bf-9e53-4069-9aa3-a4e60a3464f6 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received unexpected event network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 for instance with vm_state building and task_state spawning. [ 1015.039371] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Successfully updated port: d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.062053] env[61964]: INFO nova.compute.manager [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Took 12.97 seconds to build instance. [ 1015.175608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.565s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.177905] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.755s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.178170] env[61964]: DEBUG nova.objects.instance [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lazy-loading 'resources' on Instance uuid 47378856-4cbf-4cf6-aecd-ae935885df3d {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.198658] env[61964]: INFO nova.scheduler.client.report [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance f7f18f7b-5e58-426c-aa5d-7063a50c32ef [ 1015.542277] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.542435] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.542620] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1015.567025] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c9bc015f-c786-4f04-a22f-9ed5b352b544 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.483s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.705820] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d87698f1-aa97-4067-9e50-e1597f3b6563 tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "f7f18f7b-5e58-426c-aa5d-7063a50c32ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.279s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.807777] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0553eca7-9ac8-48cf-a9ad-f3e625db3a80 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.817693] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45774feb-2a62-4985-9ec3-08a260177121 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.848322] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e243fd-5831-46b0-8baf-04fe8b0a1471 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.856100] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1fa651-ee4b-41a9-ba26-c4e1d485e413 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.869879] env[61964]: DEBUG nova.compute.provider_tree [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.074585] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1016.209376] env[61964]: DEBUG nova.network.neutron [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.245657] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.245902] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.246149] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.246315] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.246502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.248462] env[61964]: INFO nova.compute.manager [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Terminating instance [ 1016.372646] env[61964]: DEBUG nova.scheduler.client.report [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1016.531380] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.531674] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.712541] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.712899] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance network_info: |[{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1016.713265] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:90:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd7fbf12c-2a5d-4e7f-a51e-b087b2088571', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.720658] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating folder: Project (4d63f8afe09843efb54de6ea85f276aa). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1016.720938] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-955cea6d-7c04-4d63-a566-70339a1d2453 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.731585] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created folder: Project (4d63f8afe09843efb54de6ea85f276aa) in parent group-v230360. [ 1016.731760] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating folder: Instances. Parent ref: group-v230558. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1016.731980] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8310b940-aee0-420b-bdff-61ec88d4d3d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.739823] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created folder: Instances in parent group-v230558. [ 1016.740051] env[61964]: DEBUG oslo.service.loopingcall [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.740238] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1016.740422] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef500130-0c39-497a-9dad-f453fded30ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.754319] env[61964]: DEBUG nova.compute.manager [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1016.754527] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1016.755271] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1b758e-6c73-45f3-94b3-1bdce5953293 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.762494] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.763582] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ee85ddd-87fd-40a9-9e59-cb5a6706fb21 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.764844] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.764844] env[61964]: value = "task-1041476" [ 1016.764844] env[61964]: _type = "Task" [ 1016.764844] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.768737] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1016.768737] env[61964]: value = "task-1041477" [ 1016.768737] env[61964]: _type = "Task" [ 1016.768737] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.774657] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041476, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.779271] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.877548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.881114] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.038s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.881924] env[61964]: INFO nova.compute.claims [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.901598] env[61964]: INFO nova.scheduler.client.report [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Deleted allocations for instance 47378856-4cbf-4cf6-aecd-ae935885df3d [ 1016.933890] env[61964]: DEBUG nova.compute.manager [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1016.934201] env[61964]: DEBUG nova.compute.manager [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing instance network info cache due to event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1016.934435] env[61964]: DEBUG oslo_concurrency.lockutils [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.934742] env[61964]: DEBUG oslo_concurrency.lockutils [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.934847] env[61964]: DEBUG nova.network.neutron [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.034891] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1017.277476] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041476, 'name': CreateVM_Task, 'duration_secs': 0.333302} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.277947] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1017.278644] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.278823] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.279154] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.279391] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e737b2a4-7732-4c99-8a31-66f55b71d0ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.283327] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041477, 'name': PowerOffVM_Task, 'duration_secs': 0.196091} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.283835] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.284016] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1017.284240] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-610e412f-0a50-46b8-aa9f-39a9b32073df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.286329] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1017.286329] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52123e80-a463-5e90-6d27-9fea970ce18d" [ 1017.286329] env[61964]: _type = "Task" [ 1017.286329] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.293862] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52123e80-a463-5e90-6d27-9fea970ce18d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.344783] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1017.345029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1017.345225] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleting the datastore file [datastore1] 0c1e5fc1-4e25-489f-b640-253ce03139d5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.345488] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dd6d14e-c705-47df-bff1-6db09bd9499b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.350984] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for the task: (returnval){ [ 1017.350984] env[61964]: value = "task-1041479" [ 1017.350984] env[61964]: _type = "Task" [ 1017.350984] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.358507] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.409892] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3f5542fb-ff84-43bb-93b8-92ca1f7675e5 tempest-AttachInterfacesTestJSON-278144726 tempest-AttachInterfacesTestJSON-278144726-project-member] Lock "47378856-4cbf-4cf6-aecd-ae935885df3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.357s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.555032] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.791639] env[61964]: DEBUG nova.network.neutron [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updated VIF entry in instance network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.791988] env[61964]: DEBUG nova.network.neutron [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.802104] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52123e80-a463-5e90-6d27-9fea970ce18d, 'name': SearchDatastore_Task, 'duration_secs': 0.00877} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.802104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.802104] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.802104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.802104] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.802104] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.802104] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08b89600-213d-4a79-8e41-18a7f2b4851d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.808929] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.809271] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1017.810063] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90cc41fd-03d3-406f-9b2d-bece56135b91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.815487] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1017.815487] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4372a-51b4-e2bf-08b3-6c10ced74856" [ 1017.815487] env[61964]: _type = "Task" [ 1017.815487] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.823227] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4372a-51b4-e2bf-08b3-6c10ced74856, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.860557] env[61964]: DEBUG oslo_vmware.api [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Task: {'id': task-1041479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115801} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.861351] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.861721] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1017.862046] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1017.862346] env[61964]: INFO nova.compute.manager [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1017.862688] env[61964]: DEBUG oslo.service.loopingcall [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.862997] env[61964]: DEBUG nova.compute.manager [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1017.863213] env[61964]: DEBUG nova.network.neutron [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1018.013092] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbc4adb-23bb-48cf-9b54-8e517f301e7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.019456] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1b2020-9dc3-476b-9c83-f9e9a8a0e5db {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.059433] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4a0973-eba8-4f3e-ab72-5ec188e13d89 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.064428] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f00e75-522d-46a8-b71a-b8f3a76a2b1b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.080322] env[61964]: DEBUG nova.compute.provider_tree [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.294640] env[61964]: DEBUG oslo_concurrency.lockutils [req-a5eeb3b9-a39f-4913-81c8-9d2651478169 req-d29b70ec-4838-4468-a182-142d633f7ee0 service nova] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.326690] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a4372a-51b4-e2bf-08b3-6c10ced74856, 'name': SearchDatastore_Task, 'duration_secs': 0.007609} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.326992] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a52766fa-55fd-4bd0-a1a4-f654c6c8fcae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.332693] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1018.332693] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5228e5f9-06a6-ce5a-8960-cd821437ce47" [ 1018.332693] env[61964]: _type = "Task" [ 1018.332693] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.339873] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5228e5f9-06a6-ce5a-8960-cd821437ce47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.546975] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1018.547262] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230557', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'name': 'volume-b92af457-aeb9-4914-a6b6-8291bbfc0491', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15e92512-3c92-43f0-891c-84bdaad1d186', 'attached_at': '', 'detached_at': '', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'serial': 'b92af457-aeb9-4914-a6b6-8291bbfc0491'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1018.548175] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be4b496-7e2d-470f-a8d6-5c268622d512 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.563980] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2131f78e-8b06-495d-80f2-fcd15d0c34c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.587326] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-b92af457-aeb9-4914-a6b6-8291bbfc0491/volume-b92af457-aeb9-4914-a6b6-8291bbfc0491.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.589882] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a31d0c79-f06d-4fd7-9185-ade7fb09f7ae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.607512] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1018.607512] env[61964]: value = "task-1041480" [ 1018.607512] env[61964]: _type = "Task" [ 1018.607512] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.615684] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041480, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.619873] env[61964]: ERROR nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [req-06edfa2c-94ae-4db6-947d-487ef6009675] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-06edfa2c-94ae-4db6-947d-487ef6009675"}]} [ 1018.638781] env[61964]: DEBUG nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1018.664669] env[61964]: DEBUG nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1018.664920] env[61964]: DEBUG nova.compute.provider_tree [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.679809] env[61964]: DEBUG nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1018.686081] env[61964]: DEBUG nova.network.neutron [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.702914] env[61964]: DEBUG nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1018.843774] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5228e5f9-06a6-ce5a-8960-cd821437ce47, 'name': SearchDatastore_Task, 'duration_secs': 0.043456} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.844076] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.844354] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1018.844633] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f3c8c3b-caaa-4860-8815-c40693ab9387 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.849413] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be08904c-8669-4010-b0ba-72820b8a9ba8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.853027] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1018.853027] env[61964]: value = "task-1041481" [ 1018.853027] env[61964]: _type = "Task" [ 1018.853027] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.864014] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f987781-2847-4556-9b53-6aa12763c94e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.867200] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.898567] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdc3072-a61e-49d8-844d-42879b440068 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.907157] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609e8e26-5e82-4007-b7ae-29e44c5613cf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.921445] env[61964]: DEBUG nova.compute.provider_tree [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.971301] env[61964]: DEBUG nova.compute.manager [req-7c8151fd-5f90-4201-b526-1048eab05a66 req-976fad3e-6bbc-489e-ad44-793150c7174e service nova] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Received event network-vif-deleted-e6c0d0c2-f3b5-4220-b51e-2b8ddc235e02 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1019.121865] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041480, 'name': ReconfigVM_Task, 'duration_secs': 0.340193} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.122227] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-b92af457-aeb9-4914-a6b6-8291bbfc0491/volume-b92af457-aeb9-4914-a6b6-8291bbfc0491.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.129023] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c185cf3c-ad7c-4efc-b3a1-5ad6da29ba3f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.148918] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1019.148918] env[61964]: value = "task-1041482" [ 1019.148918] env[61964]: _type = "Task" [ 1019.148918] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.158538] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041482, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.188427] env[61964]: INFO nova.compute.manager [-] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Took 1.32 seconds to deallocate network for instance. [ 1019.363948] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460781} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.364242] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1019.364460] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.364712] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16eb6528-82e2-4fdd-b89a-ddd11ed5288e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.371645] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1019.371645] env[61964]: value = "task-1041483" [ 1019.371645] env[61964]: _type = "Task" [ 1019.371645] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.381388] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.453629] env[61964]: DEBUG nova.scheduler.client.report [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1019.453629] env[61964]: DEBUG nova.compute.provider_tree [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 133 to 134 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1019.454399] env[61964]: DEBUG nova.compute.provider_tree [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.658802] env[61964]: DEBUG oslo_vmware.api [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041482, 'name': ReconfigVM_Task, 'duration_secs': 0.182724} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.659239] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230557', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'name': 'volume-b92af457-aeb9-4914-a6b6-8291bbfc0491', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '15e92512-3c92-43f0-891c-84bdaad1d186', 'attached_at': '', 'detached_at': '', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'serial': 'b92af457-aeb9-4914-a6b6-8291bbfc0491'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1019.696836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.883108] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057428} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.883108] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1019.883912] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d881ed77-b724-49f8-90b6-71a934b5d49d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.907900] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.908614] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aefbecf3-9133-4a00-9bfd-9ec360fe959f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.927565] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1019.927565] env[61964]: value = "task-1041484" [ 1019.927565] env[61964]: _type = "Task" [ 1019.927565] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.935961] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041484, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.959062] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.079s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.959621] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1019.962221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.407s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.963777] env[61964]: INFO nova.compute.claims [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.437249] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041484, 'name': ReconfigVM_Task, 'duration_secs': 0.25023} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.437538] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.438184] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95054dde-d28a-4f08-b154-514b737efa4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.445023] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1020.445023] env[61964]: value = "task-1041485" [ 1020.445023] env[61964]: _type = "Task" [ 1020.445023] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.452323] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041485, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.467740] env[61964]: DEBUG nova.compute.utils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.470866] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1020.471043] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.537062] env[61964]: DEBUG nova.policy [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c192cc0b1db94b3ab2d2c4e7af6a2e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6cf0c31f6ac649b48cfb3205d9456483', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.701814] env[61964]: DEBUG nova.objects.instance [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'flavor' on Instance uuid 15e92512-3c92-43f0-891c-84bdaad1d186 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.955095] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041485, 'name': Rename_Task, 'duration_secs': 0.14192} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.955413] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1020.955806] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4935691f-030d-4b5d-9832-c2495dd0e6bb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.961805] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1020.961805] env[61964]: value = "task-1041486" [ 1020.961805] env[61964]: _type = "Task" [ 1020.961805] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.970939] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041486, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.976102] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1021.142888] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadbcf58-8f8e-4e40-a723-90193b7a8771 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.151621] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261a912c-cd6c-4a7b-8ea0-e2f3ad554e00 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.188186] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4549f053-66a2-43a5-8d8d-09ec33e9e051 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.196465] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421a3d5e-63c0-473e-9a71-72f560624fff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.212591] env[61964]: DEBUG nova.compute.provider_tree [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.213913] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d624ec80-7a3c-4cae-b124-b384cd17b9a2 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.276s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.333964] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Successfully created port: bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.471739] env[61964]: DEBUG oslo_vmware.api [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041486, 'name': PowerOnVM_Task, 'duration_secs': 0.433441} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.472036] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1021.472268] env[61964]: INFO nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Took 6.84 seconds to spawn the instance on the hypervisor. [ 1021.472454] env[61964]: DEBUG nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1021.473235] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d615e420-1e56-43dc-87a1-6aa73a427573 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.736707] env[61964]: ERROR nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [req-f18d0cd4-297e-4171-8277-a8a5e58e772a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f18d0cd4-297e-4171-8277-a8a5e58e772a"}]} [ 1021.762352] env[61964]: DEBUG nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1021.778342] env[61964]: DEBUG nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1021.778550] env[61964]: DEBUG nova.compute.provider_tree [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1021.793879] env[61964]: DEBUG nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1021.815662] env[61964]: DEBUG nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1021.978632] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f84d3b0-6cd0-4e30-93fc-d052f7a8bbc9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.992488] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1021.995873] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14877537-330a-4d4b-a648-fcd2301fbb17 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.999302] env[61964]: INFO nova.compute.manager [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Took 11.62 seconds to build instance. [ 1022.030404] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582ab360-b3a7-4040-b3d1-b69c2b669768 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.039603] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1022.039841] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.040012] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1022.040201] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.040354] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1022.040536] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1022.040707] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1022.040865] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1022.041048] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1022.041240] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1022.041430] env[61964]: DEBUG nova.virt.hardware [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1022.042667] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfdcc71-69bf-45da-bf2b-cc87354feb5b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.047053] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc13bbce-7d24-4a5a-8992-4a147293b622 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.061613] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b14febe-b7ea-42f5-a719-58c801e7d01b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.067658] env[61964]: DEBUG nova.compute.provider_tree [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.329787] env[61964]: DEBUG nova.compute.manager [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Stashing vm_state: active {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 1022.502525] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24517170-0bb5-4adf-b4d8-7d648f95af34 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.134s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.612628] env[61964]: DEBUG nova.scheduler.client.report [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1022.613219] env[61964]: DEBUG nova.compute.provider_tree [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 136 to 137 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1022.613431] env[61964]: DEBUG nova.compute.provider_tree [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1022.852664] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.986301] env[61964]: DEBUG nova.compute.manager [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1022.986566] env[61964]: DEBUG nova.compute.manager [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing instance network info cache due to event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1022.986732] env[61964]: DEBUG oslo_concurrency.lockutils [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.986975] env[61964]: DEBUG oslo_concurrency.lockutils [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.988888] env[61964]: DEBUG nova.network.neutron [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.122252] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.160s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.122795] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1023.128976] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.432s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.129225] env[61964]: DEBUG nova.objects.instance [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lazy-loading 'resources' on Instance uuid 0c1e5fc1-4e25-489f-b640-253ce03139d5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.191383] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Successfully updated port: bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.263712] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "90679b1f-2c9c-4049-bfb9-db573220163b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.266092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.266092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.266092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.266092] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.267630] env[61964]: INFO nova.compute.manager [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Terminating instance [ 1023.630504] env[61964]: DEBUG nova.compute.utils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1023.631883] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1023.632058] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1023.678831] env[61964]: DEBUG nova.policy [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb7caa7a433e4945a82b2c7294f39f4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb92e5b578d54f1499b00aa08e7841c2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1023.696767] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.696951] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.697103] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.773444] env[61964]: DEBUG nova.compute.manager [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1023.773655] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1023.775048] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cc399e-1ff5-4c5e-9002-6190ff8e0fd3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.781358] env[61964]: DEBUG nova.network.neutron [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updated VIF entry in instance network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.781358] env[61964]: DEBUG nova.network.neutron [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.785041] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32dbda5-57c2-4b0b-a587-c28a2df411f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.789371] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.789982] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57678a92-ad2e-4e0a-b511-340d44f7766d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.794507] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07c6673-0606-4dd1-925f-c4cd2aeb52c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.799492] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1023.799492] env[61964]: value = "task-1041487" [ 1023.799492] env[61964]: _type = "Task" [ 1023.799492] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.830047] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9b587c-18da-43a0-9d50-6d76243533bf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.836102] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.841084] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3b8efc-7176-4276-a235-832f6f3efa1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.855373] env[61964]: DEBUG nova.compute.provider_tree [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.947065] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Successfully created port: 139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.050023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.050023] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.137151] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1024.238543] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1024.284862] env[61964]: DEBUG oslo_concurrency.lockutils [req-2b1cf681-d7c6-4926-b424-5f3ed1db7a89 req-df0cf573-4305-43a2-a5ca-d18e71c06451 service nova] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.311534] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041487, 'name': PowerOffVM_Task, 'duration_secs': 0.186372} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.314272] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.314525] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.314993] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cee497e-d13c-4ad4-b477-344d53950432 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.358957] env[61964]: DEBUG nova.scheduler.client.report [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1024.384400] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.384632] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.384817] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleting the datastore file [datastore2] 90679b1f-2c9c-4049-bfb9-db573220163b {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.388109] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-475bdc30-c69f-487c-8e6e-c5ade01ddf85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.400431] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1024.400431] env[61964]: value = "task-1041489" [ 1024.400431] env[61964]: _type = "Task" [ 1024.400431] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.410120] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.435460] env[61964]: DEBUG nova.network.neutron [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Updating instance_info_cache with network_info: [{"id": "bb810889-9be9-478e-afb8-c0948b0e947d", "address": "fa:16:3e:55:f4:3b", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb810889-9b", "ovs_interfaceid": "bb810889-9be9-478e-afb8-c0948b0e947d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.549457] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1024.864614] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.735s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.867660] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.015s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.889367] env[61964]: INFO nova.scheduler.client.report [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Deleted allocations for instance 0c1e5fc1-4e25-489f-b640-253ce03139d5 [ 1024.911607] env[61964]: DEBUG oslo_vmware.api [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14449} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.911881] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.912091] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1024.912279] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1024.912456] env[61964]: INFO nova.compute.manager [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1024.912697] env[61964]: DEBUG oslo.service.loopingcall [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.912911] env[61964]: DEBUG nova.compute.manager [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1024.913027] env[61964]: DEBUG nova.network.neutron [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.937955] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.938299] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Instance network_info: |[{"id": "bb810889-9be9-478e-afb8-c0948b0e947d", "address": "fa:16:3e:55:f4:3b", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb810889-9b", "ovs_interfaceid": "bb810889-9be9-478e-afb8-c0948b0e947d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1024.938714] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:f4:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4cb37d4-2060-48b6-9e60-156a71fc7ee3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb810889-9be9-478e-afb8-c0948b0e947d', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.946110] env[61964]: DEBUG oslo.service.loopingcall [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.946557] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.946790] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e328820-a2f7-466e-bddb-2604219d1b7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.968651] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.968651] env[61964]: value = "task-1041490" [ 1024.968651] env[61964]: _type = "Task" [ 1024.968651] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.978785] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041490, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.013596] env[61964]: DEBUG nova.compute.manager [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Received event network-vif-plugged-bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1025.013773] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Acquiring lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.013991] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.014183] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.014356] env[61964]: DEBUG nova.compute.manager [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] No waiting events found dispatching network-vif-plugged-bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1025.014683] env[61964]: WARNING nova.compute.manager [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Received unexpected event network-vif-plugged-bb810889-9be9-478e-afb8-c0948b0e947d for instance with vm_state building and task_state spawning. [ 1025.014683] env[61964]: DEBUG nova.compute.manager [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Received event network-changed-bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1025.014831] env[61964]: DEBUG nova.compute.manager [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Refreshing instance network info cache due to event network-changed-bb810889-9be9-478e-afb8-c0948b0e947d. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1025.015113] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Acquiring lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.015321] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Acquired lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.015493] env[61964]: DEBUG nova.network.neutron [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Refreshing network info cache for port bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1025.078380] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.147300] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1025.184835] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1025.185106] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.185273] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1025.185462] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.185646] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1025.185760] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1025.185969] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1025.186158] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1025.186332] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1025.186496] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1025.186669] env[61964]: DEBUG nova.virt.hardware [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1025.187567] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ad220b-6dc7-4282-9710-f6090409265a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.195574] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c45298-7bc9-45a1-8a3d-e0f6cbdb06de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.375758] env[61964]: INFO nova.compute.claims [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.398180] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4d8ffd64-8b28-4618-be9b-4fa946de4b87 tempest-ServerDiskConfigTestJSON-630730986 tempest-ServerDiskConfigTestJSON-630730986-project-member] Lock "0c1e5fc1-4e25-489f-b640-253ce03139d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.152s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.456042] env[61964]: DEBUG nova.compute.manager [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Received event network-vif-plugged-139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1025.456042] env[61964]: DEBUG oslo_concurrency.lockutils [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.456042] env[61964]: DEBUG oslo_concurrency.lockutils [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.456042] env[61964]: DEBUG oslo_concurrency.lockutils [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.456042] env[61964]: DEBUG nova.compute.manager [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] No waiting events found dispatching network-vif-plugged-139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1025.456042] env[61964]: WARNING nova.compute.manager [req-2408c4dd-b2c1-4c21-ae8e-8312f9d37945 req-a248fc27-28df-4416-b4cd-21a656291231 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Received unexpected event network-vif-plugged-139c3706-9f15-4801-af8d-879960814bd9 for instance with vm_state building and task_state spawning. [ 1025.480587] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041490, 'name': CreateVM_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.604406] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Successfully updated port: 139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.737976] env[61964]: DEBUG nova.network.neutron [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Updated VIF entry in instance network info cache for port bb810889-9be9-478e-afb8-c0948b0e947d. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.738377] env[61964]: DEBUG nova.network.neutron [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Updating instance_info_cache with network_info: [{"id": "bb810889-9be9-478e-afb8-c0948b0e947d", "address": "fa:16:3e:55:f4:3b", "network": {"id": "3de684c1-02cf-4188-8346-48d5fb9fa2ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-1321733004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6cf0c31f6ac649b48cfb3205d9456483", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4cb37d4-2060-48b6-9e60-156a71fc7ee3", "external-id": "nsx-vlan-transportzone-819", "segmentation_id": 819, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb810889-9b", "ovs_interfaceid": "bb810889-9be9-478e-afb8-c0948b0e947d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.883340] env[61964]: INFO nova.compute.resource_tracker [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating resource usage from migration 855a8b82-6235-430a-81b3-7ec9e263272f [ 1025.900226] env[61964]: DEBUG nova.network.neutron [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.983519] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041490, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.038317] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e70d9a8-a059-483c-b5f5-deb5cd6fe245 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.046329] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d0fd38-e903-45f5-9822-0ced6156c015 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.078804] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff592db2-6a50-43cf-bc82-f1978d18eec4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.086517] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb33fadf-284c-430f-ac18-068d1da0c516 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.100802] env[61964]: DEBUG nova.compute.provider_tree [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.107574] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.107773] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.107865] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1026.244722] env[61964]: DEBUG oslo_concurrency.lockutils [req-9c0f40f0-61e2-4f46-8d39-52e30808299a req-0dd8ba7e-ca49-4307-9661-b82f12b12576 service nova] Releasing lock "refresh_cache-3dc30299-8462-4714-8e1c-8a10412e84d5" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.403529] env[61964]: INFO nova.compute.manager [-] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Took 1.49 seconds to deallocate network for instance. [ 1026.479923] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041490, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.622883] env[61964]: ERROR nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [req-60b612f7-e06b-4d32-8f5a-06ff2ffb2a98] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-60b612f7-e06b-4d32-8f5a-06ff2ffb2a98"}]} [ 1026.639315] env[61964]: DEBUG nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1026.641770] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1026.659249] env[61964]: DEBUG nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1026.659523] env[61964]: DEBUG nova.compute.provider_tree [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.672414] env[61964]: DEBUG nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1026.690052] env[61964]: DEBUG nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1026.827857] env[61964]: DEBUG nova.network.neutron [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.845550] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0082cf-6176-42dc-a26f-14f835a6d62a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.854441] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c47c92-8d29-4806-8d30-97c1f416c6ef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.887789] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a890e649-6065-44de-8c18-5b56fc6279fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.895370] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9312b9e4-e633-4dd3-a0d7-2e85ccfd76e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.910618] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.911162] env[61964]: DEBUG nova.compute.provider_tree [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.983666] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041490, 'name': CreateVM_Task, 'duration_secs': 1.618547} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.983845] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1026.984741] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.985085] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.985448] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.986538] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b609fcb-b20e-411e-8ae1-73dc33ea7819 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.991025] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1026.991025] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52108506-8714-9523-7aeb-48f9365f896d" [ 1026.991025] env[61964]: _type = "Task" [ 1026.991025] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.998525] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52108506-8714-9523-7aeb-48f9365f896d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.039096] env[61964]: DEBUG nova.compute.manager [req-f9fcc806-a31c-4f1b-b1ef-2594cc93a8f7 req-1c0d1aa6-bf53-4d4c-945f-bce74c17a88f service nova] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Received event network-vif-deleted-f0e7e755-ae40-4d0d-bf57-2706d53067d6 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1027.335546] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.335903] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Instance network_info: |[{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1027.336419] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:a2:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '139c3706-9f15-4801-af8d-879960814bd9', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.344325] env[61964]: DEBUG oslo.service.loopingcall [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.344586] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1027.344843] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75a8f841-41ba-4619-81ab-8dc8bff8ab8a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.367021] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.367021] env[61964]: value = "task-1041491" [ 1027.367021] env[61964]: _type = "Task" [ 1027.367021] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.374018] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041491, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.444875] env[61964]: DEBUG nova.scheduler.client.report [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1027.445223] env[61964]: DEBUG nova.compute.provider_tree [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 138 to 139 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1027.445435] env[61964]: DEBUG nova.compute.provider_tree [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.487899] env[61964]: DEBUG nova.compute.manager [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Received event network-changed-139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1027.488189] env[61964]: DEBUG nova.compute.manager [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Refreshing instance network info cache due to event network-changed-139c3706-9f15-4801-af8d-879960814bd9. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1027.488433] env[61964]: DEBUG oslo_concurrency.lockutils [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] Acquiring lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.488578] env[61964]: DEBUG oslo_concurrency.lockutils [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] Acquired lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.488740] env[61964]: DEBUG nova.network.neutron [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Refreshing network info cache for port 139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.502104] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52108506-8714-9523-7aeb-48f9365f896d, 'name': SearchDatastore_Task, 'duration_secs': 0.009989} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.503025] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.503300] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.503583] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.503767] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.504010] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.504570] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f12ce7d6-a5c8-450b-8151-b59b01f55cdf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.513765] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.514268] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1027.515128] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5b69981-5afa-4901-9ee5-5616b00be06c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.520124] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1027.520124] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52343f69-c228-cd79-9464-585d609d0519" [ 1027.520124] env[61964]: _type = "Task" [ 1027.520124] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.530098] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52343f69-c228-cd79-9464-585d609d0519, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.875014] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041491, 'name': CreateVM_Task, 'duration_secs': 0.321285} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.875150] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1027.875895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.876155] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.876571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.876836] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e81f9d9d-5646-4582-9d23-3aa63ae2d7ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.881379] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1027.881379] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200fcc0-86bc-4dae-0600-056a720e079f" [ 1027.881379] env[61964]: _type = "Task" [ 1027.881379] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.889649] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200fcc0-86bc-4dae-0600-056a720e079f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.951039] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.083s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.951192] env[61964]: INFO nova.compute.manager [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Migrating [ 1027.957798] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.880s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.959248] env[61964]: INFO nova.compute.claims [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.038066] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52343f69-c228-cd79-9464-585d609d0519, 'name': SearchDatastore_Task, 'duration_secs': 0.008371} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.039161] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90f1e94f-e6b7-4d44-a84a-c26ff1ed9e3a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.046289] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1028.046289] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b219be-4a60-438d-51b5-0a148a78da3f" [ 1028.046289] env[61964]: _type = "Task" [ 1028.046289] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.060463] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b219be-4a60-438d-51b5-0a148a78da3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.270813] env[61964]: DEBUG nova.network.neutron [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updated VIF entry in instance network info cache for port 139c3706-9f15-4801-af8d-879960814bd9. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.270813] env[61964]: DEBUG nova.network.neutron [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.392491] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5200fcc0-86bc-4dae-0600-056a720e079f, 'name': SearchDatastore_Task, 'duration_secs': 0.009399} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.392799] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.393055] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1028.393302] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.393455] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.393635] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1028.393896] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7d39ff5-317a-4a19-82c0-578ba69694d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.402074] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.402368] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1028.403438] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e6027a3-3294-47ea-a5df-5fe02641f83f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.408459] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1028.408459] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a6447-1abe-2743-6df3-1b19ab805328" [ 1028.408459] env[61964]: _type = "Task" [ 1028.408459] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.416834] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a6447-1abe-2743-6df3-1b19ab805328, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.471837] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.472015] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.472238] env[61964]: DEBUG nova.network.neutron [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.556157] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b219be-4a60-438d-51b5-0a148a78da3f, 'name': SearchDatastore_Task, 'duration_secs': 0.009439} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.556616] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.556733] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 3dc30299-8462-4714-8e1c-8a10412e84d5/3dc30299-8462-4714-8e1c-8a10412e84d5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1028.557075] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9fe7afc-d869-41be-916f-753f0b553801 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.563782] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1028.563782] env[61964]: value = "task-1041492" [ 1028.563782] env[61964]: _type = "Task" [ 1028.563782] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.570767] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.774793] env[61964]: DEBUG oslo_concurrency.lockutils [req-b937e794-6b84-4714-bf51-4603bdbba154 req-b851bd0b-1655-44e9-b4eb-708f68070cbd service nova] Releasing lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.920702] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527a6447-1abe-2743-6df3-1b19ab805328, 'name': SearchDatastore_Task, 'duration_secs': 0.008681} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.921671] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c86c9ef3-c70e-49b0-bd2a-555c76211890 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.928364] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1028.928364] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb58b5-6dbd-f363-ac14-a209e8b62bea" [ 1028.928364] env[61964]: _type = "Task" [ 1028.928364] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.938974] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb58b5-6dbd-f363-ac14-a209e8b62bea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.079148] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041492, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490777} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.079148] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 3dc30299-8462-4714-8e1c-8a10412e84d5/3dc30299-8462-4714-8e1c-8a10412e84d5.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1029.079148] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.079148] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-822c4aec-1411-417b-af71-b8e3a8d50e6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.091037] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1029.091037] env[61964]: value = "task-1041493" [ 1029.091037] env[61964]: _type = "Task" [ 1029.091037] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.100487] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041493, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.137917] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be33fbb-5e4c-4a92-8432-5cae416ba1b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.147427] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5009aec-508d-4246-bf82-69b2041e8bb8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.183217] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7152a83e-1ffa-4ffc-b783-171cd11656b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.190674] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c290679-f31f-4e80-a4a4-2fe14ddcb471 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.205638] env[61964]: DEBUG nova.compute.provider_tree [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.398992] env[61964]: DEBUG nova.network.neutron [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.438681] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52cb58b5-6dbd-f363-ac14-a209e8b62bea, 'name': SearchDatastore_Task, 'duration_secs': 0.026958} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.438937] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.439235] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1029.439491] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13683594-2b48-4380-bd87-746b9ae2238b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.445868] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1029.445868] env[61964]: value = "task-1041494" [ 1029.445868] env[61964]: _type = "Task" [ 1029.445868] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.453485] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041494, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.601340] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041493, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066032} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.601638] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.602445] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e385c3-5983-42d6-8047-0180717badbc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.626511] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 3dc30299-8462-4714-8e1c-8a10412e84d5/3dc30299-8462-4714-8e1c-8a10412e84d5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.626924] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0aa40f3-877b-46cf-ab10-b48df84b950a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.649210] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1029.649210] env[61964]: value = "task-1041495" [ 1029.649210] env[61964]: _type = "Task" [ 1029.649210] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.657495] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.709378] env[61964]: DEBUG nova.scheduler.client.report [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1029.902541] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.960068] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041494, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476976} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.960068] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1029.960068] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.960068] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ecc15e4-8269-484a-a8f3-bc25cb791e18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.967019] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1029.967019] env[61964]: value = "task-1041496" [ 1029.967019] env[61964]: _type = "Task" [ 1029.967019] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.974409] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041496, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.160458] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041495, 'name': ReconfigVM_Task, 'duration_secs': 0.316893} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.160676] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 3dc30299-8462-4714-8e1c-8a10412e84d5/3dc30299-8462-4714-8e1c-8a10412e84d5.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.161337] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25515ecf-9397-49fe-8d2d-3bbed362a6b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.172141] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1030.172141] env[61964]: value = "task-1041497" [ 1030.172141] env[61964]: _type = "Task" [ 1030.172141] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.181224] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041497, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.214361] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.214853] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1030.217496] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.307s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.217735] env[61964]: DEBUG nova.objects.instance [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'resources' on Instance uuid 90679b1f-2c9c-4049-bfb9-db573220163b {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.476012] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041496, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062044} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.476665] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.477620] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fcc7805-a075-492b-bf71-b9e7092ea51d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.503015] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.503374] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d0341ae-94bd-4c9e-bb45-488295b92809 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.522276] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1030.522276] env[61964]: value = "task-1041498" [ 1030.522276] env[61964]: _type = "Task" [ 1030.522276] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.530356] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041498, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.682717] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041497, 'name': Rename_Task, 'duration_secs': 0.177407} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.683028] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1030.683286] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bae7d510-2240-4a56-8040-703cf9b1d84b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.689860] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1030.689860] env[61964]: value = "task-1041499" [ 1030.689860] env[61964]: _type = "Task" [ 1030.689860] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.698091] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041499, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.720530] env[61964]: DEBUG nova.compute.utils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.722122] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1030.722354] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1030.768396] env[61964]: DEBUG nova.policy [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '964164ab8ff54ac6b978ba27c7b62bb4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19cc478ccac141eb92c298e44b0dfa15', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.882253] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b27d93-6e0e-4e61-bed9-965c4b085515 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.889267] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc79ad2a-22c8-435d-97b2-6c9d7013922a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.925912] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26715900-9e7b-4239-bbe8-5571a576a559 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.931827] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13a5bb6-aaa5-403f-963b-8a1657c06ede {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.947833] env[61964]: DEBUG nova.compute.provider_tree [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.037596] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041498, 'name': ReconfigVM_Task, 'duration_secs': 0.302602} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.038109] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfigured VM instance instance-00000066 to attach disk [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.038779] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a639bb10-c523-4599-b959-46c6f22262fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.046490] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1031.046490] env[61964]: value = "task-1041500" [ 1031.046490] env[61964]: _type = "Task" [ 1031.046490] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.055475] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041500, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.188338] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Successfully created port: f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.203561] env[61964]: DEBUG oslo_vmware.api [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041499, 'name': PowerOnVM_Task, 'duration_secs': 0.491849} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.203897] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1031.204174] env[61964]: INFO nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Took 9.21 seconds to spawn the instance on the hypervisor. [ 1031.204636] env[61964]: DEBUG nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.205534] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7f8e9f-a680-4e7c-8fa5-d5a474d30268 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.228016] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1031.431466] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379ab858-277e-4ba7-a018-d3282f9c08e5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.454718] env[61964]: DEBUG nova.scheduler.client.report [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1031.459296] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1031.556726] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041500, 'name': Rename_Task, 'duration_secs': 0.175964} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.557154] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1031.557336] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd5fefaa-dfcf-416c-b6e1-2276d0585e96 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.563845] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1031.563845] env[61964]: value = "task-1041501" [ 1031.563845] env[61964]: _type = "Task" [ 1031.563845] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.577083] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.725409] env[61964]: INFO nova.compute.manager [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Took 19.90 seconds to build instance. [ 1031.966809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.969959] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.970628] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f26f3eb-c298-477c-b1ac-d009bc57fc3b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.978725] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1031.978725] env[61964]: value = "task-1041502" [ 1031.978725] env[61964]: _type = "Task" [ 1031.978725] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.987418] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.988510] env[61964]: INFO nova.scheduler.client.report [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted allocations for instance 90679b1f-2c9c-4049-bfb9-db573220163b [ 1032.080738] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "204fc3b1-c4a6-43ef-a408-e41be883034e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.081068] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.082847] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041501, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.109146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.109383] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.140441] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "9b48b942-5e29-4fea-83b7-a1551f78345c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.140841] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.227840] env[61964]: DEBUG oslo_concurrency.lockutils [None req-922cd303-9781-40f5-9066-b8761ab42fed tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.407s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.237469] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1032.265544] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.265821] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.265982] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.266187] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.266346] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.266500] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.266710] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.266880] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.267103] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.267286] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.267462] env[61964]: DEBUG nova.virt.hardware [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1032.268343] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7595d6-30a6-485e-8346-d72ead3ecac3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.276785] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c453cf5c-9d7f-43ed-8676-a9ad0a51ae87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.488059] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041502, 'name': PowerOffVM_Task, 'duration_secs': 0.252832} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.488361] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.488549] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1032.499457] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "3dc30299-8462-4714-8e1c-8a10412e84d5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.499640] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.499811] env[61964]: DEBUG nova.compute.manager [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.500479] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5ee507d6-d63a-465d-9c9c-80038a298602 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "90679b1f-2c9c-4049-bfb9-db573220163b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.236s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.503875] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e6f7c9-9309-421f-b032-47c9ed12bb7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.508564] env[61964]: DEBUG nova.compute.manager [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61964) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1032.509165] env[61964]: DEBUG nova.objects.instance [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'flavor' on Instance uuid 3dc30299-8462-4714-8e1c-8a10412e84d5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.575565] env[61964]: DEBUG oslo_vmware.api [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041501, 'name': PowerOnVM_Task, 'duration_secs': 0.550275} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.575565] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1032.575565] env[61964]: INFO nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Took 7.43 seconds to spawn the instance on the hypervisor. [ 1032.575565] env[61964]: DEBUG nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.575565] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e946663-8a92-4cfd-b04d-18796f70367a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.584026] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1032.611548] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1032.645425] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1032.915886] env[61964]: DEBUG nova.compute.manager [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Received event network-vif-plugged-f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1032.915886] env[61964]: DEBUG oslo_concurrency.lockutils [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] Acquiring lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.915886] env[61964]: DEBUG oslo_concurrency.lockutils [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.915886] env[61964]: DEBUG oslo_concurrency.lockutils [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.915886] env[61964]: DEBUG nova.compute.manager [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] No waiting events found dispatching network-vif-plugged-f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1032.916769] env[61964]: WARNING nova.compute.manager [req-987cc3b4-cf8f-4519-bfc2-618bb03e8010 req-34a06724-0b2d-4617-833e-fef84c08035a service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Received unexpected event network-vif-plugged-f9c93de9-1684-47f5-9fc5-7d665865ad58 for instance with vm_state building and task_state spawning. [ 1032.974787] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Successfully updated port: f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.995767] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.996151] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.996338] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.996530] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.996720] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.996872] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.997158] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.997334] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.997514] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.997667] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.998033] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1033.003643] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a82e2417-7ea9-4e9f-9e4b-4fb2fd66b968 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.024371] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1033.024371] env[61964]: value = "task-1041503" [ 1033.024371] env[61964]: _type = "Task" [ 1033.024371] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.033600] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041503, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.097883] env[61964]: INFO nova.compute.manager [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Took 15.56 seconds to build instance. [ 1033.110781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.111057] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.112654] env[61964]: INFO nova.compute.claims [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.136186] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.168430] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.476059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.476059] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquired lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.476237] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.523406] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.523759] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5385fdd1-c29b-4e7e-9aa1-de05fbccb50d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.534432] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041503, 'name': ReconfigVM_Task, 'duration_secs': 0.326727} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.535729] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1033.539107] env[61964]: DEBUG oslo_vmware.api [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1033.539107] env[61964]: value = "task-1041504" [ 1033.539107] env[61964]: _type = "Task" [ 1033.539107] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.550720] env[61964]: DEBUG oslo_vmware.api [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.600120] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7e883f2d-b046-4d7e-ae0a-f1026b5b8e7b tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.068s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.025500] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1034.041812] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1034.042096] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.042268] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1034.042456] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.042604] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1034.042754] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1034.042961] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1034.043139] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1034.043310] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1034.043481] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1034.043659] env[61964]: DEBUG nova.virt.hardware [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1034.049403] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1034.051711] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b83f6c4-2cac-4806-bdf8-292ffcaf8c4d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.067257] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.067485] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.076115] env[61964]: DEBUG oslo_vmware.api [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041504, 'name': PowerOffVM_Task, 'duration_secs': 0.352217} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.077724] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1034.077724] env[61964]: DEBUG nova.compute.manager [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1034.077858] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1034.077858] env[61964]: value = "task-1041505" [ 1034.077858] env[61964]: _type = "Task" [ 1034.077858] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.079048] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2608d73-0e66-445d-b77d-89327bc00a6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.096077] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.146397] env[61964]: DEBUG nova.compute.manager [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Stashing vm_state: active {{(pid=61964) _prep_resize /opt/stack/nova/nova/compute/manager.py:6059}} [ 1034.196181] env[61964]: DEBUG nova.network.neutron [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updating instance_info_cache with network_info: [{"id": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "address": "fa:16:3e:ef:ab:ca", "network": {"id": "4bb2d9a5-744d-450f-b085-02c8573cdc33", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-500891801-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19cc478ccac141eb92c298e44b0dfa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9c93de9-16", "ovs_interfaceid": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.292227] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4a8220-6083-434d-9d7c-32e1bdcfa10c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.300654] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8ec19a-74a3-48e7-96d6-559dfdf05a7f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.330339] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2b5e7c-96cd-4f16-9979-45883360da5a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.337422] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f17868-c038-42f8-8d39-98669dcde474 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.351565] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.572801] env[61964]: INFO nova.compute.manager [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Detaching volume 9930fa82-359c-40a3-9658-852e13df340f [ 1034.591760] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041505, 'name': ReconfigVM_Task, 'duration_secs': 0.197973} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.592100] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1034.592896] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78af1c6d-7acf-48f2-9fd6-e82bc6a0811a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.619682] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.620395] env[61964]: DEBUG oslo_concurrency.lockutils [None req-f60b437e-9fe7-490b-ac24-46a858320f47 tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.121s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.622249] env[61964]: INFO nova.virt.block_device [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Attempting to driver detach volume 9930fa82-359c-40a3-9658-852e13df340f from mountpoint /dev/sdb [ 1034.622493] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1034.622714] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230545', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'name': 'volume-9930fa82-359c-40a3-9658-852e13df340f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a14232f-5cc6-4856-9291-eb0a1bd2ab69', 'attached_at': '', 'detached_at': '', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'serial': '9930fa82-359c-40a3-9658-852e13df340f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1034.623330] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fc09391-eb8e-4096-81de-24f50cb4149f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.637439] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284712b0-d6a0-4fd6-b324-274ce67b26d7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.662970] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b1c425-99e6-4445-b3da-739d6e61bd4e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.666200] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1034.666200] env[61964]: value = "task-1041506" [ 1034.666200] env[61964]: _type = "Task" [ 1034.666200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.672502] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.673790] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a90c95-4fff-4be9-b320-13402ec723c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.679477] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.697252] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58654c9c-e19e-4fb8-9fdb-9faf6f3a4cae {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.701914] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Releasing lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.701914] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Instance network_info: |[{"id": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "address": "fa:16:3e:ef:ab:ca", "network": {"id": "4bb2d9a5-744d-450f-b085-02c8573cdc33", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-500891801-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19cc478ccac141eb92c298e44b0dfa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9c93de9-16", "ovs_interfaceid": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1034.712601] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:ab:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9c93de9-1684-47f5-9fc5-7d665865ad58', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.720278] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Creating folder: Project (19cc478ccac141eb92c298e44b0dfa15). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1034.720752] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] The volume has not been displaced from its original location: [datastore1] volume-9930fa82-359c-40a3-9658-852e13df340f/volume-9930fa82-359c-40a3-9658-852e13df340f.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1034.726589] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfiguring VM instance instance-00000053 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1034.727059] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3cbdb26-af9a-4a73-8155-dac4cffbee80 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.728736] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4df48aaa-fc3f-4922-bdf6-8ddd6889e79c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.752843] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1034.752843] env[61964]: value = "task-1041508" [ 1034.752843] env[61964]: _type = "Task" [ 1034.752843] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.757531] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Created folder: Project (19cc478ccac141eb92c298e44b0dfa15) in parent group-v230360. [ 1034.757750] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Creating folder: Instances. Parent ref: group-v230563. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1034.758347] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d4e6458-65be-4787-a751-558e8f9866f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.763812] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041508, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.771848] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Created folder: Instances in parent group-v230563. [ 1034.772156] env[61964]: DEBUG oslo.service.loopingcall [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.772388] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1034.772690] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5db6593-fdfa-4033-bfd6-f14da1a0a053 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.792654] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.792654] env[61964]: value = "task-1041510" [ 1034.792654] env[61964]: _type = "Task" [ 1034.792654] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.800842] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041510, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.854806] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1034.941954] env[61964]: DEBUG nova.compute.manager [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Received event network-changed-f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1034.942150] env[61964]: DEBUG nova.compute.manager [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Refreshing instance network info cache due to event network-changed-f9c93de9-1684-47f5-9fc5-7d665865ad58. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1034.942414] env[61964]: DEBUG oslo_concurrency.lockutils [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] Acquiring lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.942613] env[61964]: DEBUG oslo_concurrency.lockutils [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] Acquired lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.942907] env[61964]: DEBUG nova.network.neutron [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Refreshing network info cache for port f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.175309] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041506, 'name': ReconfigVM_Task, 'duration_secs': 0.321748} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.175595] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.175881] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1035.248955] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "3dc30299-8462-4714-8e1c-8a10412e84d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.249302] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.249525] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.249716] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.249908] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.252468] env[61964]: INFO nova.compute.manager [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Terminating instance [ 1035.264175] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041508, 'name': ReconfigVM_Task, 'duration_secs': 0.249664} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.264319] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Reconfigured VM instance instance-00000053 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1035.270554] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cea426c-4ef6-46af-953d-ae06f2a66ba8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.286237] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1035.286237] env[61964]: value = "task-1041511" [ 1035.286237] env[61964]: _type = "Task" [ 1035.286237] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.294865] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041511, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.302891] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041510, 'name': CreateVM_Task, 'duration_secs': 0.348854} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.303163] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1035.303878] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.304058] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.304385] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.304627] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30baf7cd-002b-428d-a0d9-fdb8fa7d5098 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.308909] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1035.308909] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524fbbde-ccaf-61da-2f0c-10645c8bd2ec" [ 1035.308909] env[61964]: _type = "Task" [ 1035.308909] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.317900] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524fbbde-ccaf-61da-2f0c-10645c8bd2ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.361400] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.361893] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1035.364358] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.228s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.365725] env[61964]: INFO nova.compute.claims [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.679418] env[61964]: DEBUG nova.network.neutron [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updated VIF entry in instance network info cache for port f9c93de9-1684-47f5-9fc5-7d665865ad58. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1035.679789] env[61964]: DEBUG nova.network.neutron [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updating instance_info_cache with network_info: [{"id": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "address": "fa:16:3e:ef:ab:ca", "network": {"id": "4bb2d9a5-744d-450f-b085-02c8573cdc33", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-500891801-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19cc478ccac141eb92c298e44b0dfa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9c93de9-16", "ovs_interfaceid": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.682973] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49bc046-59f3-455c-badb-8a22ed7fa75f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.704497] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc967dc9-3c29-4eb4-8560-7c22fa436afd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.724365] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1035.764777] env[61964]: DEBUG nova.compute.manager [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1035.765046] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1035.765877] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f9959b-6a73-44a7-978e-ebec135b0665 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.773523] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.773754] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd6e77a8-6528-4711-8215-4afa191c5304 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.795047] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.818031] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]524fbbde-ccaf-61da-2f0c-10645c8bd2ec, 'name': SearchDatastore_Task, 'duration_secs': 0.010977} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.818408] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.818653] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.818903] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.819063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.819253] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.819528] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9de7e64a-a23d-4925-9a39-79b4b5430ed2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.827907] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.828123] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1035.829168] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18451e41-3ad9-4109-9f8c-d4244a485fde {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.834388] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1035.834388] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52841958-b8a3-8390-7091-49f262cb3705" [ 1035.834388] env[61964]: _type = "Task" [ 1035.834388] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.838595] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1035.838824] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1035.839040] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore1] 3dc30299-8462-4714-8e1c-8a10412e84d5 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.839647] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5544e8c7-beda-423d-ad86-ed6c8208bda0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.844357] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52841958-b8a3-8390-7091-49f262cb3705, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.848254] env[61964]: DEBUG oslo_vmware.api [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1035.848254] env[61964]: value = "task-1041513" [ 1035.848254] env[61964]: _type = "Task" [ 1035.848254] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.855705] env[61964]: DEBUG oslo_vmware.api [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.870333] env[61964]: DEBUG nova.compute.utils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.873914] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1035.874110] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1035.915370] env[61964]: DEBUG nova.policy [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '247c47d6cd4a48e288aa20ce4e0693c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76df35828e24e55b42182df9b12d3d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.167302] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Successfully created port: 02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.183335] env[61964]: DEBUG oslo_concurrency.lockutils [req-bc37bf58-4c85-4a71-8b38-bceff01698e6 req-f6529759-2660-4f67-b60c-82bf45fd9fc6 service nova] Releasing lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.299481] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041511, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.345363] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52841958-b8a3-8390-7091-49f262cb3705, 'name': SearchDatastore_Task, 'duration_secs': 0.034305} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.346188] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6369442f-c22f-4db0-83a0-6c20cbcb3ef4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.352928] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1036.352928] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52780def-08e0-369a-f850-ff7b9492b28e" [ 1036.352928] env[61964]: _type = "Task" [ 1036.352928] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.359191] env[61964]: DEBUG oslo_vmware.api [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209033} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.359832] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.360111] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1036.360339] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1036.360531] env[61964]: INFO nova.compute.manager [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1036.360790] env[61964]: DEBUG oslo.service.loopingcall [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.360997] env[61964]: DEBUG nova.compute.manager [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1036.361168] env[61964]: DEBUG nova.network.neutron [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1036.365341] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52780def-08e0-369a-f850-ff7b9492b28e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.381379] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1036.580248] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39815c7-66fb-4f4b-a73e-be7ed8f40086 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.587560] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0680e237-b44d-4e76-b9d3-0195b40d9a73 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.617725] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b52e307-52b5-4b3a-b4c3-967547ccc45e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.624779] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3fb228-2e0e-45e0-b99a-aaff434d0bc0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.637851] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.660825] env[61964]: DEBUG nova.compute.manager [req-87f55aae-8b25-4c61-8f5b-35a0e9b8c412 req-4e6c218f-e248-450f-ac0f-de70cb6e1ece service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Received event network-vif-deleted-bb810889-9be9-478e-afb8-c0948b0e947d {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1036.660825] env[61964]: INFO nova.compute.manager [req-87f55aae-8b25-4c61-8f5b-35a0e9b8c412 req-4e6c218f-e248-450f-ac0f-de70cb6e1ece service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Neutron deleted interface bb810889-9be9-478e-afb8-c0948b0e947d; detaching it from the instance and deleting it from the info cache [ 1036.660825] env[61964]: DEBUG nova.network.neutron [req-87f55aae-8b25-4c61-8f5b-35a0e9b8c412 req-4e6c218f-e248-450f-ac0f-de70cb6e1ece service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.798313] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041511, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.864192] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52780def-08e0-369a-f850-ff7b9492b28e, 'name': SearchDatastore_Task, 'duration_secs': 0.030773} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.864511] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.864806] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ee273d32-a49f-43b3-8b08-8017eb1b1636/ee273d32-a49f-43b3-8b08-8017eb1b1636.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1036.865100] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6ab7d9b-ecc7-432e-97c2-95a18b0f5f6d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.871250] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1036.871250] env[61964]: value = "task-1041514" [ 1036.871250] env[61964]: _type = "Task" [ 1036.871250] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.879211] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.141943] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1037.145488] env[61964]: DEBUG nova.network.neutron [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.164822] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11160325-3353-4cbc-8e3c-e57ae64b5fbf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.175653] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e4c6cb-5b43-4ff6-9295-0d059031579b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.207016] env[61964]: DEBUG nova.compute.manager [req-87f55aae-8b25-4c61-8f5b-35a0e9b8c412 req-4e6c218f-e248-450f-ac0f-de70cb6e1ece service nova] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Detach interface failed, port_id=bb810889-9be9-478e-afb8-c0948b0e947d, reason: Instance 3dc30299-8462-4714-8e1c-8a10412e84d5 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1037.298847] env[61964]: DEBUG oslo_vmware.api [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041511, 'name': ReconfigVM_Task, 'duration_secs': 1.792058} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.299182] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230545', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'name': 'volume-9930fa82-359c-40a3-9658-852e13df340f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8a14232f-5cc6-4856-9291-eb0a1bd2ab69', 'attached_at': '', 'detached_at': '', 'volume_id': '9930fa82-359c-40a3-9658-852e13df340f', 'serial': '9930fa82-359c-40a3-9658-852e13df340f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1037.369862] env[61964]: DEBUG nova.network.neutron [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Port da9bd08d-0a86-4d20-82b3-0160a3cf26b2 binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1037.382198] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460307} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.382496] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] ee273d32-a49f-43b3-8b08-8017eb1b1636/ee273d32-a49f-43b3-8b08-8017eb1b1636.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1037.382720] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.383024] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14e4ab12-d53c-494f-9558-421c4c921ae2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.390338] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1037.390338] env[61964]: value = "task-1041515" [ 1037.390338] env[61964]: _type = "Task" [ 1037.390338] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.394414] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1037.401453] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.420828] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1037.421106] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.421309] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1037.421507] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.421659] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1037.421814] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1037.422042] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1037.422212] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1037.422385] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1037.422549] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1037.422725] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1037.423872] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bada86-6c38-4dc5-9691-233eb47ca6a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.432884] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2534140-a551-45d9-ae39-db402e02e26e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.647876] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.648437] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1037.651031] env[61964]: INFO nova.compute.manager [-] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Took 1.29 seconds to deallocate network for instance. [ 1037.651536] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.483s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.652892] env[61964]: INFO nova.compute.claims [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.732637] env[61964]: DEBUG nova.compute.manager [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Received event network-vif-plugged-02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1037.732866] env[61964]: DEBUG oslo_concurrency.lockutils [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] Acquiring lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.733095] env[61964]: DEBUG oslo_concurrency.lockutils [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.733274] env[61964]: DEBUG oslo_concurrency.lockutils [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.733447] env[61964]: DEBUG nova.compute.manager [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] No waiting events found dispatching network-vif-plugged-02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1037.733619] env[61964]: WARNING nova.compute.manager [req-441c34be-03cc-4654-8df5-7643c197b17d req-39fd586d-7307-4e5d-a507-07131cb25da6 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Received unexpected event network-vif-plugged-02d68f9a-04ba-4510-adde-65b7a27eafdd for instance with vm_state building and task_state spawning. [ 1037.846563] env[61964]: DEBUG nova.objects.instance [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.848578] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Successfully updated port: 02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.900901] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.241545} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.901169] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.901934] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b05a2f-d758-48b1-82be-3b42772fdd82 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.923545] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] ee273d32-a49f-43b3-8b08-8017eb1b1636/ee273d32-a49f-43b3-8b08-8017eb1b1636.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.924508] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af61beae-5d83-4692-afba-837f1ddfa32f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.944509] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1037.944509] env[61964]: value = "task-1041516" [ 1037.944509] env[61964]: _type = "Task" [ 1037.944509] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.952874] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041516, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.153276] env[61964]: DEBUG nova.compute.utils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.155032] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1038.155212] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1038.160008] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.200815] env[61964]: DEBUG nova.policy [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '247c47d6cd4a48e288aa20ce4e0693c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76df35828e24e55b42182df9b12d3d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.350191] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.350319] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.350446] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.390100] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.390335] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.390513] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.438466] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Successfully created port: 9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1038.454170] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041516, 'name': ReconfigVM_Task, 'duration_secs': 0.274082} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.454464] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Reconfigured VM instance instance-00000067 to attach disk [datastore2] ee273d32-a49f-43b3-8b08-8017eb1b1636/ee273d32-a49f-43b3-8b08-8017eb1b1636.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.455531] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b608b95-0539-4514-83bc-8c8ef2da2b35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.461675] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1038.461675] env[61964]: value = "task-1041517" [ 1038.461675] env[61964]: _type = "Task" [ 1038.461675] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.469987] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041517, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.658330] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1038.819638] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fb3139-61e5-4607-bb7e-84a6c70e185f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.827265] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17e06d8-3c7a-4d36-b464-12f6cf05ba86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.859322] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c77e5fcb-47cf-4ad8-81d9-8b998ba56e29 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.792s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.861873] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b9120b-23a1-48d7-be5e-cb715f3a25b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.872653] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d1743a-1128-45b8-b952-61b6fbf8ddb6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.889305] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1038.919301] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1038.971533] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041517, 'name': Rename_Task, 'duration_secs': 0.132277} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.973985] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.974374] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0177aa33-a56b-4acc-b215-325cdb9ecd7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.980700] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1038.980700] env[61964]: value = "task-1041518" [ 1038.980700] env[61964]: _type = "Task" [ 1038.980700] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.988467] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.109229] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Updating instance_info_cache with network_info: [{"id": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "address": "fa:16:3e:b4:ed:56", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02d68f9a-04", "ovs_interfaceid": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.368430] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.368758] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.368977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.369184] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.369358] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.371533] env[61964]: INFO nova.compute.manager [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Terminating instance [ 1039.409328] env[61964]: ERROR nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [req-de1facc4-70c9-440b-ba34-367ddbd55103] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-de1facc4-70c9-440b-ba34-367ddbd55103"}]} [ 1039.426385] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1039.428895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.429163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.429450] env[61964]: DEBUG nova.network.neutron [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.441866] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1039.442087] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1039.455202] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1039.472577] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1039.491134] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041518, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.603443] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53452ab-8476-4e85-9a36-2d11b83a8684 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.611898] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2baa214-c12c-4368-99b3-9927a8f8d9d6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.615011] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.615302] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Instance network_info: |[{"id": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "address": "fa:16:3e:b4:ed:56", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02d68f9a-04", "ovs_interfaceid": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1039.615667] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:ed:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02d68f9a-04ba-4510-adde-65b7a27eafdd', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.623070] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Creating folder: Project (b76df35828e24e55b42182df9b12d3d3). Parent ref: group-v230360. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1039.623625] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbcc90b0-6d60-435a-a2a6-c8f014881fb8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.652739] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af4c975-0bef-4598-8836-6f8953e61441 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.655135] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Created folder: Project (b76df35828e24e55b42182df9b12d3d3) in parent group-v230360. [ 1039.655320] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Creating folder: Instances. Parent ref: group-v230566. {{(pid=61964) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1039.655536] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a84ccc3a-2a07-4309-a2e7-04cc1112b3c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.661586] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867ad59c-00f9-4d5d-855c-b4e810c8a810 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.665991] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Created folder: Instances in parent group-v230566. [ 1039.666231] env[61964]: DEBUG oslo.service.loopingcall [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.666831] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1039.667053] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1e3e3a6-55cc-453f-924d-424a5a2e1a2f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.689220] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1039.691269] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1039.696340] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.696340] env[61964]: value = "task-1041521" [ 1039.696340] env[61964]: _type = "Task" [ 1039.696340] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.705213] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041521, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.715713] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1039.715979] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1039.716160] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1039.716346] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1039.716496] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1039.716661] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1039.716910] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1039.717108] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1039.717285] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1039.717451] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1039.717627] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1039.718436] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b75b5f-1266-4e78-93f9-1f74d916686b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.726057] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936a64f6-266d-46e8-b6f2-148e79ed13b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.765345] env[61964]: DEBUG nova.compute.manager [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Received event network-changed-02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1039.765558] env[61964]: DEBUG nova.compute.manager [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Refreshing instance network info cache due to event network-changed-02d68f9a-04ba-4510-adde-65b7a27eafdd. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1039.765783] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] Acquiring lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.765937] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] Acquired lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.766130] env[61964]: DEBUG nova.network.neutron [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Refreshing network info cache for port 02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.878023] env[61964]: DEBUG nova.compute.manager [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1039.878023] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1039.878023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd82c90-de87-4d4f-aaae-97488fcaed76 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.886275] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1039.886275] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c944e085-0a60-4f8d-a48a-de72cec2347f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.892489] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1039.892489] env[61964]: value = "task-1041522" [ 1039.892489] env[61964]: _type = "Task" [ 1039.892489] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.900939] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.945743] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Successfully updated port: 9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.993794] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041518, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.190344] env[61964]: DEBUG nova.network.neutron [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.207569] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041521, 'name': CreateVM_Task, 'duration_secs': 0.359332} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.207772] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1040.208695] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.208882] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.209232] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.209499] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fa70537-ef0e-49ed-8e7f-53ef57c28840 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.214472] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1040.214472] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5296a9b6-1ba9-e7cb-0734-a624c518d036" [ 1040.214472] env[61964]: _type = "Task" [ 1040.214472] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.224551] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5296a9b6-1ba9-e7cb-0734-a624c518d036, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.228837] env[61964]: DEBUG nova.scheduler.client.report [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 144 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1040.229106] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 144 to 145 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1040.229292] env[61964]: DEBUG nova.compute.provider_tree [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1040.402597] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041522, 'name': PowerOffVM_Task, 'duration_secs': 0.201599} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.405225] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.405406] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.405665] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e19de1ca-1d87-47d6-85f6-34c5d13e3563 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.447314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.447505] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.447754] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1040.475083] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.475488] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.475651] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleting the datastore file [datastore2] 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.475797] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e51ffe5f-ca49-491f-8da2-d7ad25d6f288 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.478893] env[61964]: DEBUG nova.network.neutron [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Updated VIF entry in instance network info cache for port 02d68f9a-04ba-4510-adde-65b7a27eafdd. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.479277] env[61964]: DEBUG nova.network.neutron [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Updating instance_info_cache with network_info: [{"id": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "address": "fa:16:3e:b4:ed:56", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02d68f9a-04", "ovs_interfaceid": "02d68f9a-04ba-4510-adde-65b7a27eafdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.488098] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1040.488098] env[61964]: value = "task-1041524" [ 1040.488098] env[61964]: _type = "Task" [ 1040.488098] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.495991] env[61964]: DEBUG oslo_vmware.api [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041518, 'name': PowerOnVM_Task, 'duration_secs': 1.453368} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.496707] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1040.496967] env[61964]: INFO nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Took 8.26 seconds to spawn the instance on the hypervisor. [ 1040.497181] env[61964]: DEBUG nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1040.497957] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c7ba0d-326b-4e61-9266-91d8ef75b135 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.503456] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.691850] env[61964]: DEBUG oslo_concurrency.lockutils [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.726396] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5296a9b6-1ba9-e7cb-0734-a624c518d036, 'name': SearchDatastore_Task, 'duration_secs': 0.010587} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.726810] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.727060] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.727308] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.727459] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.727639] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.727905] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1b41766-f741-490b-9cfe-83f94633102d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.733682] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.082s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.734250] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1040.737898] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.065s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.739118] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.739290] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1040.740852] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b66196ca-6780-499f-a540-d0986e6c8b19 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.747347] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1040.747347] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526ab3dc-9f8c-e43e-cc8a-dc5fbd3f9c40" [ 1040.747347] env[61964]: _type = "Task" [ 1040.747347] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.756758] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526ab3dc-9f8c-e43e-cc8a-dc5fbd3f9c40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.980185] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.982491] env[61964]: DEBUG oslo_concurrency.lockutils [req-2ec56045-ceb6-4bd5-a6dc-8172aff4c7c8 req-84f45cd0-85ea-49f2-8f8f-a13bd63dae05 service nova] Releasing lock "refresh_cache-204fc3b1-c4a6-43ef-a408-e41be883034e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.998492] env[61964]: DEBUG oslo_vmware.api [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184663} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.000790] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.000954] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.001148] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.001326] env[61964]: INFO nova.compute.manager [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1041.001566] env[61964]: DEBUG oslo.service.loopingcall [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.001794] env[61964]: DEBUG nova.compute.manager [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1041.001899] env[61964]: DEBUG nova.network.neutron [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1041.019485] env[61964]: INFO nova.compute.manager [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Took 15.96 seconds to build instance. [ 1041.118954] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Updating instance_info_cache with network_info: [{"id": "9afcca07-e043-46d7-b521-2fa0062b8e87", "address": "fa:16:3e:77:49:03", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcca07-e0", "ovs_interfaceid": "9afcca07-e043-46d7-b521-2fa0062b8e87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.201283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7274934b-b902-4923-8bc9-b13641219467 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.209729] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0b840f-08e2-4385-b524-1898e5ea7add {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.244277] env[61964]: INFO nova.compute.claims [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.248751] env[61964]: DEBUG nova.compute.utils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.250793] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1041.250793] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.014354] env[61964]: DEBUG oslo_concurrency.lockutils [None req-24312a0d-4381-4db7-923f-d08042508c0a tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.967s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.014672] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.014755] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Instance network_info: |[{"id": "9afcca07-e043-46d7-b521-2fa0062b8e87", "address": "fa:16:3e:77:49:03", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcca07-e0", "ovs_interfaceid": "9afcca07-e043-46d7-b521-2fa0062b8e87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1042.016870] env[61964]: INFO nova.compute.resource_tracker [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating resource usage from migration 90eac758-d932-4603-9dd4-78a03c6fccac [ 1042.018807] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1042.028025] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Received event network-vif-plugged-9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1042.028025] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Acquiring lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.028025] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.028025] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.028025] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] No waiting events found dispatching network-vif-plugged-9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1042.028025] env[61964]: WARNING nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Received unexpected event network-vif-plugged-9afcca07-e043-46d7-b521-2fa0062b8e87 for instance with vm_state building and task_state spawning. [ 1042.028025] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Received event network-changed-9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1042.028025] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Refreshing instance network info cache due to event network-changed-9afcca07-e043-46d7-b521-2fa0062b8e87. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1042.028025] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Acquiring lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.028025] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Acquired lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.028025] env[61964]: DEBUG nova.network.neutron [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Refreshing network info cache for port 9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.029181] env[61964]: DEBUG nova.network.neutron [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.031172] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:49:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9afcca07-e043-46d7-b521-2fa0062b8e87', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.039264] env[61964]: DEBUG oslo.service.loopingcall [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.044075] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1042.044075] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d085122-bed9-42db-996f-b8afa93a974b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.063498] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526ab3dc-9f8c-e43e-cc8a-dc5fbd3f9c40, 'name': SearchDatastore_Task, 'duration_secs': 0.021382} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.069383] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f87fabb7-764a-4089-b41f-b0614efaa0be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.075973] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.075973] env[61964]: value = "task-1041525" [ 1042.075973] env[61964]: _type = "Task" [ 1042.075973] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.082497] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1042.082497] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52436083-a86f-f17b-e339-cc32ac603b4e" [ 1042.082497] env[61964]: _type = "Task" [ 1042.082497] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.091318] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041525, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.095684] env[61964]: DEBUG nova.policy [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '247c47d6cd4a48e288aa20ce4e0693c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76df35828e24e55b42182df9b12d3d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.104068] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52436083-a86f-f17b-e339-cc32ac603b4e, 'name': SearchDatastore_Task, 'duration_secs': 0.010485} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.104393] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.104730] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 204fc3b1-c4a6-43ef-a408-e41be883034e/204fc3b1-c4a6-43ef-a408-e41be883034e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1042.105110] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f0129cc-da5e-4300-9f9c-2b1dd3f54b2e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.118982] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1042.118982] env[61964]: value = "task-1041526" [ 1042.118982] env[61964]: _type = "Task" [ 1042.118982] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.134228] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.287382] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c16d9a-5770-4280-b9f0-e5a52da9a63e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.296682] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f97ba5-81e4-41ad-b03c-48c0d926f62e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.355159] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e28955-e136-4cff-aa4e-bb88bde707b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.368956] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f3e1f1-c3cd-47ce-9227-3fdc3df15850 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.386718] env[61964]: DEBUG nova.compute.provider_tree [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.481443] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Successfully created port: f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.530446] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.530848] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.531213] env[61964]: INFO nova.compute.manager [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Rebooting instance [ 1042.543213] env[61964]: INFO nova.compute.manager [-] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Took 1.54 seconds to deallocate network for instance. [ 1042.594022] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041525, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.596160] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadc0235-03ab-4d4c-9192-f635803f982b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.620701] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8847cf1-5842-4022-b8fb-7c3c8995ce98 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.632441] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494885} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.634784] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 204fc3b1-c4a6-43ef-a408-e41be883034e/204fc3b1-c4a6-43ef-a408-e41be883034e.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1042.635037] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.635395] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1042.641654] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa34459d-b241-403a-a2bb-5f7c2e12e75f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.650582] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1042.650582] env[61964]: value = "task-1041527" [ 1042.650582] env[61964]: _type = "Task" [ 1042.650582] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.661538] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.825020] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.825287] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.836727] env[61964]: DEBUG nova.network.neutron [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Updated VIF entry in instance network info cache for port 9afcca07-e043-46d7-b521-2fa0062b8e87. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1042.837071] env[61964]: DEBUG nova.network.neutron [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Updating instance_info_cache with network_info: [{"id": "9afcca07-e043-46d7-b521-2fa0062b8e87", "address": "fa:16:3e:77:49:03", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9afcca07-e0", "ovs_interfaceid": "9afcca07-e043-46d7-b521-2fa0062b8e87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.888885] env[61964]: DEBUG nova.scheduler.client.report [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1043.043918] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1043.052335] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.054745] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.054916] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquired lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.055122] env[61964]: DEBUG nova.network.neutron [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.070187] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1043.070450] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.070614] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1043.070800] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.070953] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1043.071121] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1043.071331] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1043.071493] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1043.071659] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1043.071822] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1043.071993] env[61964]: DEBUG nova.virt.hardware [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1043.073096] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644052c7-2221-40da-bfc4-e96e3396ec08 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.082070] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475805da-1e2e-4e9c-885b-70413965cc07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.104084] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041525, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.143737] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1043.144037] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36392921-0401-40f0-b8ed-924af76db74c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.155106] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1043.155106] env[61964]: value = "task-1041528" [ 1043.155106] env[61964]: _type = "Task" [ 1043.155106] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.161027] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063065} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.161653] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.162406] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d52bc0-8ac8-4f91-b4f4-06c786eaa2f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.167546] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041528, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.188181] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 204fc3b1-c4a6-43ef-a408-e41be883034e/204fc3b1-c4a6-43ef-a408-e41be883034e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.188555] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f602377e-df20-43b4-9900-3e9da1542abf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.210430] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1043.210430] env[61964]: value = "task-1041529" [ 1043.210430] env[61964]: _type = "Task" [ 1043.210430] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.219209] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041529, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.328939] env[61964]: DEBUG nova.compute.utils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.339159] env[61964]: DEBUG oslo_concurrency.lockutils [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] Releasing lock "refresh_cache-6ad71a14-c3d3-4db3-968a-d51d51d86496" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.339417] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Received event network-vif-deleted-d9618c13-facd-40f7-b201-776d03dffe57 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1043.339593] env[61964]: INFO nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Neutron deleted interface d9618c13-facd-40f7-b201-776d03dffe57; detaching it from the instance and deleting it from the info cache [ 1043.339769] env[61964]: DEBUG nova.network.neutron [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.393284] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.655s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.393518] env[61964]: INFO nova.compute.manager [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Migrating [ 1043.400430] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.240s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.400686] env[61964]: DEBUG nova.objects.instance [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid 3dc30299-8462-4714-8e1c-8a10412e84d5 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.595943] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4617b326-9c61-4fdc-a548-61b1cd7b2268 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.601796] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041525, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.607067] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a258d0db-d507-4438-bf7b-4e44925055dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.640783] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149ee0be-9db7-4d39-aeac-c04cf937d8a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.649073] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131acf74-e5b2-4626-befa-c298a1d9ad7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.663506] env[61964]: DEBUG nova.compute.provider_tree [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.678438] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041528, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.721764] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041529, 'name': ReconfigVM_Task, 'duration_secs': 0.318914} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.722076] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 204fc3b1-c4a6-43ef-a408-e41be883034e/204fc3b1-c4a6-43ef-a408-e41be883034e.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.722718] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71d712a9-fa52-40b3-8446-701f4777a494 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.729507] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1043.729507] env[61964]: value = "task-1041530" [ 1043.729507] env[61964]: _type = "Task" [ 1043.729507] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.738018] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041530, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.824495] env[61964]: DEBUG nova.network.neutron [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updating instance_info_cache with network_info: [{"id": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "address": "fa:16:3e:ef:ab:ca", "network": {"id": "4bb2d9a5-744d-450f-b085-02c8573cdc33", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-500891801-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19cc478ccac141eb92c298e44b0dfa15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9c93de9-16", "ovs_interfaceid": "f9c93de9-1684-47f5-9fc5-7d665865ad58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.831791] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.845064] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ed80500-d1f7-4435-aee3-7fe79ae82304 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.853645] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef65dfa-826c-4336-8a4b-35e296c2581b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.885480] env[61964]: DEBUG nova.compute.manager [req-ba145799-a6c8-44b7-ad88-bfc4668dc025 req-97eef93f-0e57-46c3-8027-7858ee0024a6 service nova] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Detach interface failed, port_id=d9618c13-facd-40f7-b201-776d03dffe57, reason: Instance 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1043.915016] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.915237] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.915421] env[61964]: DEBUG nova.network.neutron [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1043.929903] env[61964]: DEBUG nova.compute.manager [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Received event network-vif-plugged-f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1043.930135] env[61964]: DEBUG oslo_concurrency.lockutils [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] Acquiring lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.930362] env[61964]: DEBUG oslo_concurrency.lockutils [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.930620] env[61964]: DEBUG oslo_concurrency.lockutils [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.930674] env[61964]: DEBUG nova.compute.manager [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] No waiting events found dispatching network-vif-plugged-f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1043.930841] env[61964]: WARNING nova.compute.manager [req-9f50a010-305c-4793-b98d-ef7345a3fe3b req-3b1aa097-bed1-407c-b138-bbbb08f13f0d service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Received unexpected event network-vif-plugged-f0603f94-32f6-46df-a2fe-498227b76538 for instance with vm_state building and task_state spawning. [ 1044.019780] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Successfully updated port: f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.097714] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041525, 'name': CreateVM_Task, 'duration_secs': 1.542966} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.098029] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1044.098621] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.098804] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.099149] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.099413] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f7c9bda-c84d-458c-ab47-3d6720d7b71d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.104577] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1044.104577] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c39da3-0748-38c7-1d54-def8f6944447" [ 1044.104577] env[61964]: _type = "Task" [ 1044.104577] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.112881] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c39da3-0748-38c7-1d54-def8f6944447, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.171135] env[61964]: DEBUG nova.scheduler.client.report [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.177667] env[61964]: DEBUG oslo_vmware.api [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041528, 'name': PowerOnVM_Task, 'duration_secs': 0.836644} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.178249] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1044.178461] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-df7b5b83-4c63-4602-8818-113587654b14 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance '15e92512-3c92-43f0-891c-84bdaad1d186' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1044.240837] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041530, 'name': Rename_Task, 'duration_secs': 0.164945} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.241130] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1044.241375] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cac6c7c1-8b78-44a2-b280-72df9d48498e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.248294] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1044.248294] env[61964]: value = "task-1041531" [ 1044.248294] env[61964]: _type = "Task" [ 1044.248294] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.257293] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.327956] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Releasing lock "refresh_cache-ee273d32-a49f-43b3-8b08-8017eb1b1636" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.523029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.523242] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.523368] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.620165] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52c39da3-0748-38c7-1d54-def8f6944447, 'name': SearchDatastore_Task, 'duration_secs': 0.019221} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.620548] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.620831] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.621101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.621247] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.621426] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.621691] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e3b4746-bf11-4d22-96d8-21f9ca4648c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.632488] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.632798] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1044.633776] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44d97008-cfd3-4931-976c-7d96dcb513ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.640728] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1044.640728] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52261cba-3bfe-fcd4-c2a0-c6da30090311" [ 1044.640728] env[61964]: _type = "Task" [ 1044.640728] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.651846] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52261cba-3bfe-fcd4-c2a0-c6da30090311, 'name': SearchDatastore_Task, 'duration_secs': 0.009257} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.652707] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b81f94a0-fb20-42d0-8a24-8a0f922794ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.661699] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1044.661699] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db3ea1-a25b-d2dc-8c89-07126ba0deeb" [ 1044.661699] env[61964]: _type = "Task" [ 1044.661699] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.671055] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db3ea1-a25b-d2dc-8c89-07126ba0deeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.679231] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.279s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.682229] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.630s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.682586] env[61964]: DEBUG nova.objects.instance [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'resources' on Instance uuid 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.707370] env[61964]: INFO nova.scheduler.client.report [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance 3dc30299-8462-4714-8e1c-8a10412e84d5 [ 1044.714631] env[61964]: DEBUG nova.network.neutron [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.759228] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041531, 'name': PowerOnVM_Task, 'duration_secs': 0.490746} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.759564] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1044.759711] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1044.759891] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1044.760879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfe5c5c-431e-4e25-a76f-8f10f247936c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.833402] env[61964]: DEBUG nova.compute.manager [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1044.834573] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fb642a-faea-4b33-b3c1-37855092dd4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.896352] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.896822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.897127] env[61964]: INFO nova.compute.manager [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attaching volume 63d77f81-da29-4cef-a84c-71585e99374b to /dev/sdb [ 1044.928069] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a130abee-965f-4c31-bb2c-8c67fe75a618 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.935766] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe6ceec-4e0d-423f-965e-90536ec465a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.949934] env[61964]: DEBUG nova.virt.block_device [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating existing volume attachment record: cc36dfaf-41b1-4121-85e8-f7a5879f478a {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1045.054037] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.173732] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52db3ea1-a25b-d2dc-8c89-07126ba0deeb, 'name': SearchDatastore_Task, 'duration_secs': 0.010437} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.174707] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.174707] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 6ad71a14-c3d3-4db3-968a-d51d51d86496/6ad71a14-c3d3-4db3-968a-d51d51d86496.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1045.174967] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84c6bfd1-b678-444c-ba8c-3c4603131f67 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.183262] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1045.183262] env[61964]: value = "task-1041533" [ 1045.183262] env[61964]: _type = "Task" [ 1045.183262] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.195815] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.200961] env[61964]: DEBUG nova.network.neutron [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Updating instance_info_cache with network_info: [{"id": "f0603f94-32f6-46df-a2fe-498227b76538", "address": "fa:16:3e:f3:61:1d", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0603f94-32", "ovs_interfaceid": "f0603f94-32f6-46df-a2fe-498227b76538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.216397] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d911e83c-f0be-4b28-91e1-228eee76051f tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "3dc30299-8462-4714-8e1c-8a10412e84d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.967s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.217817] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.279218] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Took 12.19 seconds to build instance. [ 1045.378872] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e640be-a84a-4f8d-a4fe-936f8f071d5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.389393] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d445861a-c5b5-42e1-9caf-3fd78522f052 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.422938] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee736a7-dfe5-490f-b8e4-4ab7ddba4f42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.432412] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5385dc4a-e781-4f1c-8629-bc60299b5d5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.461991] env[61964]: DEBUG nova.compute.provider_tree [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.694339] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461694} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.694609] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 6ad71a14-c3d3-4db3-968a-d51d51d86496/6ad71a14-c3d3-4db3-968a-d51d51d86496.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1045.694829] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.695096] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8144c2cc-4e25-4ea4-8247-ca14c06f7b58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.703469] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.703895] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Instance network_info: |[{"id": "f0603f94-32f6-46df-a2fe-498227b76538", "address": "fa:16:3e:f3:61:1d", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0603f94-32", "ovs_interfaceid": "f0603f94-32f6-46df-a2fe-498227b76538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1045.704667] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1045.704667] env[61964]: value = "task-1041534" [ 1045.704667] env[61964]: _type = "Task" [ 1045.704667] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.704904] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:61:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '975b168a-03e5-449d-95ac-4d51ba027242', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0603f94-32f6-46df-a2fe-498227b76538', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.713377] env[61964]: DEBUG oslo.service.loopingcall [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.713715] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1045.717224] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f80f171-3a09-4a4a-bdbd-7c897edb2795 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.747821] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.747821] env[61964]: value = "task-1041535" [ 1045.747821] env[61964]: _type = "Task" [ 1045.747821] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.749476] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041534, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.760398] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041535, 'name': CreateVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.781471] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.700s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.851345] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3153da0f-85d3-47ec-bf71-64debdb92f41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.860850] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Doing hard reboot of VM {{(pid=61964) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 1045.861171] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-8982dd59-9340-46c1-a8ef-c234d3d0bf42 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.869104] env[61964]: DEBUG oslo_vmware.api [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1045.869104] env[61964]: value = "task-1041536" [ 1045.869104] env[61964]: _type = "Task" [ 1045.869104] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.877538] env[61964]: DEBUG oslo_vmware.api [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041536, 'name': ResetVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.956551] env[61964]: DEBUG nova.compute.manager [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Received event network-changed-f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1045.956831] env[61964]: DEBUG nova.compute.manager [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Refreshing instance network info cache due to event network-changed-f0603f94-32f6-46df-a2fe-498227b76538. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1045.957139] env[61964]: DEBUG oslo_concurrency.lockutils [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] Acquiring lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.957347] env[61964]: DEBUG oslo_concurrency.lockutils [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] Acquired lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.957895] env[61964]: DEBUG nova.network.neutron [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Refreshing network info cache for port f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.965734] env[61964]: DEBUG nova.scheduler.client.report [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.223841] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041534, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071041} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.224216] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.224885] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd89c1d-9819-4a20-bf49-4b7b3559c482 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.250088] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 6ad71a14-c3d3-4db3-968a-d51d51d86496/6ad71a14-c3d3-4db3-968a-d51d51d86496.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.251261] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4d98a0c-00d5-4645-b917-b73c9a1fc31b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.277036] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041535, 'name': CreateVM_Task, 'duration_secs': 0.336844} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.278095] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1046.278095] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1046.278095] env[61964]: value = "task-1041537" [ 1046.278095] env[61964]: _type = "Task" [ 1046.278095] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.278557] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.278652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.278970] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.279278] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-458b23f1-85cb-4e49-96ee-72e0bf2d94c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.286599] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1046.286599] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5224ee76-3d61-9f25-bdae-8852f0ed8ab3" [ 1046.286599] env[61964]: _type = "Task" [ 1046.286599] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.289681] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041537, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.296712] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5224ee76-3d61-9f25-bdae-8852f0ed8ab3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.380077] env[61964]: DEBUG oslo_vmware.api [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041536, 'name': ResetVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.472950] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.493817] env[61964]: INFO nova.scheduler.client.report [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted allocations for instance 8a14232f-5cc6-4856-9291-eb0a1bd2ab69 [ 1046.639075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.639424] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.639788] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.639938] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.640190] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.642656] env[61964]: INFO nova.compute.manager [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Terminating instance [ 1046.713469] env[61964]: DEBUG nova.network.neutron [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Updated VIF entry in instance network info cache for port f0603f94-32f6-46df-a2fe-498227b76538. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.713898] env[61964]: DEBUG nova.network.neutron [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Updating instance_info_cache with network_info: [{"id": "f0603f94-32f6-46df-a2fe-498227b76538", "address": "fa:16:3e:f3:61:1d", "network": {"id": "5dfc14f0-5a25-48b2-91cf-1bdfa1e23a1d", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1244002571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76df35828e24e55b42182df9b12d3d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "975b168a-03e5-449d-95ac-4d51ba027242", "external-id": "nsx-vlan-transportzone-365", "segmentation_id": 365, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0603f94-32", "ovs_interfaceid": "f0603f94-32f6-46df-a2fe-498227b76538", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.754463] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48261606-5e62-4fc7-816a-788e02bcbdb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.774882] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 0 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1046.787951] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.798780] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5224ee76-3d61-9f25-bdae-8852f0ed8ab3, 'name': SearchDatastore_Task, 'duration_secs': 0.072036} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.799097] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.799335] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.799613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.799779] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.799965] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.800232] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-296a708b-f347-4fd1-8ab5-add8147a6091 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.808439] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.808688] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1046.809390] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a7ac61d-b77b-40d7-aeac-624740f44cc3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.814832] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1046.814832] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dcf039-19d9-4661-1d52-64a8b7a3014f" [ 1046.814832] env[61964]: _type = "Task" [ 1046.814832] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.823309] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dcf039-19d9-4661-1d52-64a8b7a3014f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.881356] env[61964]: DEBUG oslo_vmware.api [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041536, 'name': ResetVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.001989] env[61964]: DEBUG oslo_concurrency.lockutils [None req-57f71a5e-0dcd-4933-a714-54dd37cc2032 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "8a14232f-5cc6-4856-9291-eb0a1bd2ab69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.633s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.146435] env[61964]: DEBUG nova.compute.manager [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1047.146716] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.147599] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3361a328-0956-4d87-8d3a-fac43d80ce05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.151092] env[61964]: DEBUG nova.network.neutron [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Port da9bd08d-0a86-4d20-82b3-0160a3cf26b2 binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1047.151326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.151521] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.151697] env[61964]: DEBUG nova.network.neutron [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.162987] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.162987] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5d27974-0ab5-4208-be2d-badd54241f85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.166270] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1047.166270] env[61964]: value = "task-1041538" [ 1047.166270] env[61964]: _type = "Task" [ 1047.166270] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.177352] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041538, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.218399] env[61964]: DEBUG oslo_concurrency.lockutils [req-9fb082d3-03a1-47a3-a1ba-f58da953ecae req-bb62073a-3f52-4bdc-afc6-59023e162bd5 service nova] Releasing lock "refresh_cache-9b48b942-5e29-4fea-83b7-a1551f78345c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.281547] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.281902] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-704b8b4d-cf87-40a8-a331-ea82e212cb3e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.294113] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.295725] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1047.295725] env[61964]: value = "task-1041540" [ 1047.295725] env[61964]: _type = "Task" [ 1047.295725] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.306472] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041540, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.325749] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52dcf039-19d9-4661-1d52-64a8b7a3014f, 'name': SearchDatastore_Task, 'duration_secs': 0.01169} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.326654] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4630e327-8fc6-4b81-9276-ef6170125ebf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.335116] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1047.335116] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b89d8d-5c48-dc44-1d47-a694efd1c981" [ 1047.335116] env[61964]: _type = "Task" [ 1047.335116] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.341797] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b89d8d-5c48-dc44-1d47-a694efd1c981, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.382021] env[61964]: DEBUG oslo_vmware.api [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041536, 'name': ResetVM_Task, 'duration_secs': 1.022558} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.382385] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Did hard reboot of VM {{(pid=61964) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 1047.382546] env[61964]: DEBUG nova.compute.manager [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1047.383384] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf4a111-3542-498b-8099-b76c39765001 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.675538] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041538, 'name': PowerOffVM_Task, 'duration_secs': 0.206998} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.675804] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.675978] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.676256] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17c59f19-06fc-4ccb-9932-c3ff6c6b8baf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.758812] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1047.759060] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1047.759607] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleting the datastore file [datastore1] aef4c3a7-641a-4356-9187-ae4c082ccde9 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.759607] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c039f6e9-7d27-4568-b80b-c68aa2832c03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.766006] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for the task: (returnval){ [ 1047.766006] env[61964]: value = "task-1041542" [ 1047.766006] env[61964]: _type = "Task" [ 1047.766006] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.774116] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041542, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.791633] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041537, 'name': ReconfigVM_Task, 'duration_secs': 1.113563} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.791899] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 6ad71a14-c3d3-4db3-968a-d51d51d86496/6ad71a14-c3d3-4db3-968a-d51d51d86496.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.792521] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ab29e3a-5455-479f-a317-c3292a154689 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.799914] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1047.799914] env[61964]: value = "task-1041543" [ 1047.799914] env[61964]: _type = "Task" [ 1047.799914] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.805675] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041540, 'name': PowerOffVM_Task, 'duration_secs': 0.229955} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.808040] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.808306] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 17 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1047.815016] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041543, 'name': Rename_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.843906] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52b89d8d-5c48-dc44-1d47-a694efd1c981, 'name': SearchDatastore_Task, 'duration_secs': 0.022819} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.847899] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.847899] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 9b48b942-5e29-4fea-83b7-a1551f78345c/9b48b942-5e29-4fea-83b7-a1551f78345c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1047.847899] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb37ed51-24a8-498e-bc21-3b4a29e03d77 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.854035] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1047.854035] env[61964]: value = "task-1041544" [ 1047.854035] env[61964]: _type = "Task" [ 1047.854035] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.862898] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.903164] env[61964]: DEBUG oslo_concurrency.lockutils [None req-75ac2e73-ad70-44bc-bb35-bbaf4ce83922 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.372s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.927367] env[61964]: DEBUG nova.network.neutron [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.276314] env[61964]: DEBUG oslo_vmware.api [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Task: {'id': task-1041542, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.390172} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.276530] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.276738] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1048.276932] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.277137] env[61964]: INFO nova.compute.manager [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1048.277408] env[61964]: DEBUG oslo.service.loopingcall [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.277622] env[61964]: DEBUG nova.compute.manager [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1048.277718] env[61964]: DEBUG nova.network.neutron [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.310541] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041543, 'name': Rename_Task, 'duration_secs': 0.149563} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.310917] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.311145] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb1b46dc-d656-481d-b117-9dfc3f366412 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.319980] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:32Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1048.320277] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.320445] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1048.320638] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.320791] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1048.320956] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1048.321197] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1048.321367] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1048.321541] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1048.321705] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1048.321881] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1048.327044] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1048.327044] env[61964]: value = "task-1041545" [ 1048.327044] env[61964]: _type = "Task" [ 1048.327044] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.329500] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3842fa4b-d6a3-41a7-9f91-49a05a1bac07 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.350054] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041545, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.351744] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1048.351744] env[61964]: value = "task-1041546" [ 1048.351744] env[61964]: _type = "Task" [ 1048.351744] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.366693] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041546, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.370389] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041544, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.430995] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.669414] env[61964]: DEBUG nova.compute.manager [req-e36ac0a4-e973-4458-9bae-f70a1904b39b req-20709684-d1d0-4cb0-a987-6f503bb98bc3 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Received event network-vif-deleted-74f7cab9-da08-4246-bebd-f8b2f0ec5a3b {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1048.669787] env[61964]: INFO nova.compute.manager [req-e36ac0a4-e973-4458-9bae-f70a1904b39b req-20709684-d1d0-4cb0-a987-6f503bb98bc3 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Neutron deleted interface 74f7cab9-da08-4246-bebd-f8b2f0ec5a3b; detaching it from the instance and deleting it from the info cache [ 1048.669844] env[61964]: DEBUG nova.network.neutron [req-e36ac0a4-e973-4458-9bae-f70a1904b39b req-20709684-d1d0-4cb0-a987-6f503bb98bc3 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.850279] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041545, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.868190] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699239} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.868505] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041546, 'name': ReconfigVM_Task, 'duration_secs': 0.268923} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.868748] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] 9b48b942-5e29-4fea-83b7-a1551f78345c/9b48b942-5e29-4fea-83b7-a1551f78345c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1048.868962] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.869271] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 33 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1048.872771] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34cdb337-6383-4a9d-bad4-e8920a7e5cfb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.879810] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1048.879810] env[61964]: value = "task-1041547" [ 1048.879810] env[61964]: _type = "Task" [ 1048.879810] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.887807] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.935285] env[61964]: DEBUG nova.compute.manager [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61964) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1049.145025] env[61964]: DEBUG nova.network.neutron [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.172674] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2af4192c-9981-45b3-a1c7-80457307bc93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.182677] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade260ec-9068-4662-a37d-c1126dd1fa20 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.212515] env[61964]: DEBUG nova.compute.manager [req-e36ac0a4-e973-4458-9bae-f70a1904b39b req-20709684-d1d0-4cb0-a987-6f503bb98bc3 service nova] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Detach interface failed, port_id=74f7cab9-da08-4246-bebd-f8b2f0ec5a3b, reason: Instance aef4c3a7-641a-4356-9187-ae4c082ccde9 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1049.230698] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.230942] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.231170] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.231359] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.231532] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.233803] env[61964]: INFO nova.compute.manager [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Terminating instance [ 1049.289763] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.290022] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.351081] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041545, 'name': PowerOnVM_Task, 'duration_secs': 0.635784} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.351425] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.351565] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Took 9.66 seconds to spawn the instance on the hypervisor. [ 1049.351751] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1049.352505] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0a636f-763c-4cb6-9420-fc17645b25f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.377635] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1049.377886] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.378067] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1049.378296] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.378459] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1049.378619] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1049.378827] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1049.378990] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1049.379175] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1049.379401] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1049.379527] env[61964]: DEBUG nova.virt.hardware [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1049.384781] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1049.385145] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60b2cec5-8ace-4e83-8eb8-ec0fb00bd336 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.408023] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096843} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.409244] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.409591] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1049.409591] env[61964]: value = "task-1041548" [ 1049.409591] env[61964]: _type = "Task" [ 1049.409591] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.410491] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5386f1b9-5e2d-4c0b-9b38-e1b8f97ee3b8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.422760] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041548, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.440192] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 9b48b942-5e29-4fea-83b7-a1551f78345c/9b48b942-5e29-4fea-83b7-a1551f78345c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.440722] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3390f0a-3947-4d6c-9e32-8a6dc001659e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.459212] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1049.459212] env[61964]: value = "task-1041549" [ 1049.459212] env[61964]: _type = "Task" [ 1049.459212] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.466191] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.495151] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1049.495397] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230570', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'name': 'volume-63d77f81-da29-4cef-a84c-71585e99374b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'serial': '63d77f81-da29-4cef-a84c-71585e99374b'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1049.496260] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c9b9ef-b7a9-4b4d-a373-30d96ad2d376 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.512033] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1200cf-2559-4fc1-8382-2b1159917ff2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.535542] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-63d77f81-da29-4cef-a84c-71585e99374b/volume-63d77f81-da29-4cef-a84c-71585e99374b.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.537853] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76c9381d-8217-49ce-bc11-d44d98fa0f40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.555833] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1049.555833] env[61964]: value = "task-1041550" [ 1049.555833] env[61964]: _type = "Task" [ 1049.555833] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.563599] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.647902] env[61964]: INFO nova.compute.manager [-] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Took 1.37 seconds to deallocate network for instance. [ 1049.738507] env[61964]: DEBUG nova.compute.manager [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1049.738507] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1049.739612] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559be866-3295-42e1-96d7-5577756fcb58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.748437] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1049.748554] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-069c2ece-95d3-475a-8930-db2d1de9aafb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.755638] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1049.755638] env[61964]: value = "task-1041551" [ 1049.755638] env[61964]: _type = "Task" [ 1049.755638] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.764340] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.792188] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1049.870817] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Took 16.75 seconds to build instance. [ 1049.924034] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041548, 'name': ReconfigVM_Task, 'duration_secs': 0.480219} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.924034] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1049.924643] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b495b7b4-de44-4fda-b8b1-150ea6005436 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.949607] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.949967] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe840062-b603-4179-8890-c6421a43c75d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.972848] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041549, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.974160] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1049.974160] env[61964]: value = "task-1041552" [ 1049.974160] env[61964]: _type = "Task" [ 1049.974160] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.983143] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.051577] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.051943] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.070059] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041550, 'name': ReconfigVM_Task, 'duration_secs': 0.498781} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.070411] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-63d77f81-da29-4cef-a84c-71585e99374b/volume-63d77f81-da29-4cef-a84c-71585e99374b.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.075257] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02f477d5-d196-4ab9-ac0c-28b2f1a2ed34 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.089856] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1050.089856] env[61964]: value = "task-1041553" [ 1050.089856] env[61964]: _type = "Task" [ 1050.089856] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.097742] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041553, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.156652] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.269500] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041551, 'name': PowerOffVM_Task, 'duration_secs': 0.188935} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.269946] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1050.270285] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1050.270662] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-380b79fb-384d-4133-a407-8e58496fd780 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.316197] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.332640] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1050.332931] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1050.333171] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Deleting the datastore file [datastore2] ee273d32-a49f-43b3-8b08-8017eb1b1636 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.333445] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81a49925-f74c-4e14-b454-2cf440119017 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.340081] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for the task: (returnval){ [ 1050.340081] env[61964]: value = "task-1041555" [ 1050.340081] env[61964]: _type = "Task" [ 1050.340081] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.348113] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.372722] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.263s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.473269] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041549, 'name': ReconfigVM_Task, 'duration_secs': 0.779544} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.473550] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 9b48b942-5e29-4fea-83b7-a1551f78345c/9b48b942-5e29-4fea-83b7-a1551f78345c.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.474196] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d22823c-27c9-4840-955b-7edabe94ffad {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.484399] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041552, 'name': ReconfigVM_Task, 'duration_secs': 0.450432} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.485611] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Reconfigured VM instance instance-00000066 to attach disk [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4/ada96855-b35f-4ff5-b3ce-e4fa1baaedc4.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.485751] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 50 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1050.489392] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1050.489392] env[61964]: value = "task-1041556" [ 1050.489392] env[61964]: _type = "Task" [ 1050.489392] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.497269] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041556, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.554606] env[61964]: DEBUG nova.objects.instance [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'migration_context' on Instance uuid 15e92512-3c92-43f0-891c-84bdaad1d186 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.598909] env[61964]: DEBUG oslo_vmware.api [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041553, 'name': ReconfigVM_Task, 'duration_secs': 0.144245} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.599335] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230570', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'name': 'volume-63d77f81-da29-4cef-a84c-71585e99374b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'serial': '63d77f81-da29-4cef-a84c-71585e99374b'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1050.850790] env[61964]: DEBUG oslo_vmware.api [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Task: {'id': task-1041555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153277} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.851104] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.851351] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1050.851587] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1050.851776] env[61964]: INFO nova.compute.manager [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1050.852039] env[61964]: DEBUG oslo.service.loopingcall [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.852246] env[61964]: DEBUG nova.compute.manager [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1050.852352] env[61964]: DEBUG nova.network.neutron [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1050.993499] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cb91b3-ffd9-41df-94ef-4c537b36d66e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.005028] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041556, 'name': Rename_Task, 'duration_secs': 0.396321} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.019149] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1051.019553] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c97d3dc5-2773-4b19-808b-a321535e9a4b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.021840] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f142eb-5418-4a46-82f9-5520b6922f50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.039154] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 67 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1051.044634] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1051.044634] env[61964]: value = "task-1041557" [ 1051.044634] env[61964]: _type = "Task" [ 1051.044634] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.052483] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041557, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.123343] env[61964]: DEBUG nova.compute.manager [req-07d90e2d-44c0-4da8-9317-073bf23761b8 req-eb1ba951-7b6f-429b-abb1-034fb26bebda service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Received event network-vif-deleted-f9c93de9-1684-47f5-9fc5-7d665865ad58 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1051.123601] env[61964]: INFO nova.compute.manager [req-07d90e2d-44c0-4da8-9317-073bf23761b8 req-eb1ba951-7b6f-429b-abb1-034fb26bebda service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Neutron deleted interface f9c93de9-1684-47f5-9fc5-7d665865ad58; detaching it from the instance and deleting it from the info cache [ 1051.123825] env[61964]: DEBUG nova.network.neutron [req-07d90e2d-44c0-4da8-9317-073bf23761b8 req-eb1ba951-7b6f-429b-abb1-034fb26bebda service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.234800] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b5c837-d8c8-46e1-bf11-62a768a95a84 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.242745] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3eea57-e417-4455-9f15-95ebb49f9cb0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.274940] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60ad374-fb04-44d6-866e-e295f5242337 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.282619] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98127fe-a2ab-40a5-8ed1-2b548d804b35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.296101] env[61964]: DEBUG nova.compute.provider_tree [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.558013] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041557, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.587261] env[61964]: DEBUG nova.network.neutron [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Port 139c3706-9f15-4801-af8d-879960814bd9 binding to destination host cpu-1 is already ACTIVE {{(pid=61964) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1051.601732] env[61964]: DEBUG nova.network.neutron [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.626221] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2923be8b-2a68-4e10-b22d-957a56b2a453 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.635761] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a271d8c5-f6b2-4b85-a487-122dd36683b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.647511] env[61964]: DEBUG nova.objects.instance [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.668028] env[61964]: DEBUG nova.compute.manager [req-07d90e2d-44c0-4da8-9317-073bf23761b8 req-eb1ba951-7b6f-429b-abb1-034fb26bebda service nova] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Detach interface failed, port_id=f9c93de9-1684-47f5-9fc5-7d665865ad58, reason: Instance ee273d32-a49f-43b3-8b08-8017eb1b1636 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1051.819259] env[61964]: ERROR nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [req-d12dac05-d7ae-40e0-b531-84039e78efb1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 57b292ab-02d9-4aab-ba83-292890345a17. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d12dac05-d7ae-40e0-b531-84039e78efb1"}]} [ 1051.845132] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing inventories for resource provider 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1051.859273] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating ProviderTree inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1051.859519] env[61964]: DEBUG nova.compute.provider_tree [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 182, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.871730] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing aggregate associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, aggregates: None {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1051.890203] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Refreshing trait associations for resource provider 57b292ab-02d9-4aab-ba83-292890345a17, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=61964) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1052.047200] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62550a9d-4e2c-468f-9cd3-38e9b4e56a2d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.057926] env[61964]: DEBUG oslo_vmware.api [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041557, 'name': PowerOnVM_Task, 'duration_secs': 0.80096} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.060029] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.060227] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Took 9.02 seconds to spawn the instance on the hypervisor. [ 1052.060416] env[61964]: DEBUG nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1052.061190] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5348224e-834a-41cb-9805-273af5d52181 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.064125] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ca6b42-09f6-43ac-bb44-35c11210c8fd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.104493] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5524fe4-0854-492b-8833-65d202f625a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.107572] env[61964]: INFO nova.compute.manager [-] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Took 1.26 seconds to deallocate network for instance. [ 1052.117766] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec79421f-3c20-4559-81fb-f5ee5aa2a3ff {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.132578] env[61964]: DEBUG nova.compute.provider_tree [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.153053] env[61964]: DEBUG oslo_concurrency.lockutils [None req-4f93676a-bdc7-447c-a382-6ce7827fc83d tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.256s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.620851] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.624602] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.624831] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.625018] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.627029] env[61964]: INFO nova.compute.manager [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Took 19.47 seconds to build instance. [ 1052.662595] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updated inventory for provider 57b292ab-02d9-4aab-ba83-292890345a17 with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1052.662896] env[61964]: DEBUG nova.compute.provider_tree [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating resource provider 57b292ab-02d9-4aab-ba83-292890345a17 generation from 146 to 147 during operation: update_inventory {{(pid=61964) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1052.663131] env[61964]: DEBUG nova.compute.provider_tree [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Updating inventory in ProviderTree for provider 57b292ab-02d9-4aab-ba83-292890345a17 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1053.039878] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.040128] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.130212] env[61964]: DEBUG oslo_concurrency.lockutils [None req-71fd9b4f-ceed-4c8e-8a29-64a45dca0558 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.989s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.544603] env[61964]: DEBUG nova.compute.utils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.637593] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "204fc3b1-c4a6-43ef-a408-e41be883034e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.637952] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.638239] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.638506] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.638739] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.641185] env[61964]: INFO nova.compute.manager [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Terminating instance [ 1053.675129] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.623s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.684093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.528s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.684435] env[61964]: DEBUG nova.objects.instance [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lazy-loading 'resources' on Instance uuid aef4c3a7-641a-4356-9187-ae4c082ccde9 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.692365] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.692546] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.692729] env[61964]: DEBUG nova.network.neutron [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.048246] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.145911] env[61964]: DEBUG nova.compute.manager [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1054.146208] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1054.147171] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686af6c4-11dc-4528-8be7-77fdb6ecd709 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.155040] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1054.155285] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c98b5f6d-9f0d-4c9a-b528-5a0f4c6d8299 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.161496] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1054.161496] env[61964]: value = "task-1041558" [ 1054.161496] env[61964]: _type = "Task" [ 1054.161496] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.169146] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.339815] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d635565-c582-449d-a1fe-9155cd21f0f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.348212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d511567-6b21-457a-bb2d-280520ef04c5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.382371] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907ae17d-4083-4bc9-b598-11a819a7893c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.391073] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d05f74a-a53e-4036-84b7-b29b7eb5d9f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.405613] env[61964]: DEBUG nova.compute.provider_tree [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.473709] env[61964]: DEBUG nova.network.neutron [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.672093] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041558, 'name': PowerOffVM_Task, 'duration_secs': 0.353612} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.672438] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1054.672612] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1054.672887] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cdbb014-45b6-4053-ae55-f92f142793f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.741452] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1054.741797] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1054.742020] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleting the datastore file [datastore2] 204fc3b1-c4a6-43ef-a408-e41be883034e {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.742311] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-490c08ad-da2d-481a-9dc0-b2af21970bda {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.749254] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1054.749254] env[61964]: value = "task-1041560" [ 1054.749254] env[61964]: _type = "Task" [ 1054.749254] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.757509] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.909216] env[61964]: DEBUG nova.scheduler.client.report [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1054.976056] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.107372] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.107656] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.107908] env[61964]: INFO nova.compute.manager [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attaching volume c2808fb3-02c8-4d39-9906-9d4b44c139ff to /dev/sdc [ 1055.138415] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c014a77-e25b-4722-a002-90331944b545 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.146146] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514c23c2-2bbc-4a6a-b531-4ae7854ef3cc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.159908] env[61964]: DEBUG nova.virt.block_device [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating existing volume attachment record: cf30a3b9-f327-4044-bbfb-f9ca87c18fcc {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1055.219480] env[61964]: INFO nova.compute.manager [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Swapping old allocation on dict_keys(['57b292ab-02d9-4aab-ba83-292890345a17']) held by migration 855a8b82-6235-430a-81b3-7ec9e263272f for instance [ 1055.243032] env[61964]: DEBUG nova.scheduler.client.report [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Overwriting current allocation {'allocations': {'57b292ab-02d9-4aab-ba83-292890345a17': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 147}}, 'project_id': '9b627b25af0c4982a22c8c0bd28641a3', 'user_id': '430778d9f17e4579a81902608fb3501a', 'consumer_generation': 1} on consumer 15e92512-3c92-43f0-891c-84bdaad1d186 {{(pid=61964) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1055.258687] env[61964]: DEBUG oslo_vmware.api [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256563} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.258961] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.259200] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1055.259394] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1055.259594] env[61964]: INFO nova.compute.manager [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1055.259866] env[61964]: DEBUG oslo.service.loopingcall [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.260086] env[61964]: DEBUG nova.compute.manager [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1055.260186] env[61964]: DEBUG nova.network.neutron [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1055.324331] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.324544] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.324764] env[61964]: DEBUG nova.network.neutron [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.414793] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.418653] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.101s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.420962] env[61964]: INFO nova.compute.claims [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.434590] env[61964]: INFO nova.scheduler.client.report [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Deleted allocations for instance aef4c3a7-641a-4356-9187-ae4c082ccde9 [ 1055.512081] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6874758-7305-4085-83be-95ea379ef8b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.535297] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d36335e-fa87-475b-ab98-326c7d44ce13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.543815] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 83 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1055.549067] env[61964]: DEBUG nova.compute.manager [req-8d65dd9d-12b8-4b3e-8aba-1b34a5e6fe3f req-3c2e25b5-878b-4264-a90b-aabd72312c18 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Received event network-vif-deleted-02d68f9a-04ba-4510-adde-65b7a27eafdd {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1055.549320] env[61964]: INFO nova.compute.manager [req-8d65dd9d-12b8-4b3e-8aba-1b34a5e6fe3f req-3c2e25b5-878b-4264-a90b-aabd72312c18 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Neutron deleted interface 02d68f9a-04ba-4510-adde-65b7a27eafdd; detaching it from the instance and deleting it from the info cache [ 1055.549507] env[61964]: DEBUG nova.network.neutron [req-8d65dd9d-12b8-4b3e-8aba-1b34a5e6fe3f req-3c2e25b5-878b-4264-a90b-aabd72312c18 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.946078] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2279d606-8b41-473f-b47d-6723799ec75a tempest-ServersTestJSON-1484395192 tempest-ServersTestJSON-1484395192-project-member] Lock "aef4c3a7-641a-4356-9187-ae4c082ccde9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.307s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.015296] env[61964]: DEBUG nova.network.neutron [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.044065] env[61964]: DEBUG nova.network.neutron [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [{"id": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "address": "fa:16:3e:3d:50:ca", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda9bd08d-0a", "ovs_interfaceid": "da9bd08d-0a86-4d20-82b3-0160a3cf26b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.053351] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1056.054027] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43915dc1-db51-41ef-b4e0-caaacf4a548c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.056380] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-262f9f2e-d7e2-4e3f-a1c6-f6c8420d72f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.067200] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d003175-c488-4df1-98c5-ced1dcb82116 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.077552] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1056.077552] env[61964]: value = "task-1041562" [ 1056.077552] env[61964]: _type = "Task" [ 1056.077552] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.090670] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041562, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.099573] env[61964]: DEBUG nova.compute.manager [req-8d65dd9d-12b8-4b3e-8aba-1b34a5e6fe3f req-3c2e25b5-878b-4264-a90b-aabd72312c18 service nova] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Detach interface failed, port_id=02d68f9a-04ba-4510-adde-65b7a27eafdd, reason: Instance 204fc3b1-c4a6-43ef-a408-e41be883034e could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1056.518213] env[61964]: INFO nova.compute.manager [-] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Took 1.26 seconds to deallocate network for instance. [ 1056.548964] env[61964]: DEBUG oslo_concurrency.lockutils [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-15e92512-3c92-43f0-891c-84bdaad1d186" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.551071] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114e019f-802e-4c6d-b9e9-1ce96b08f439 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.558860] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb8e079-2a56-46e1-8e9a-e0805b1d14f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.581078] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9505b9ef-64fb-41e8-ba2a-14315b5f7521 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.596077] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f0c266-cd51-41e6-b8e6-a2ab0ad2720b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.599896] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041562, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.627451] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f80f1a6-289d-40fe-8370-b9aab8bbf715 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.636997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8ed01f-3b74-4db9-9e6a-e0a0229d210f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.661071] env[61964]: DEBUG nova.compute.provider_tree [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.027094] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.090514] env[61964]: DEBUG oslo_vmware.api [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041562, 'name': PowerOnVM_Task, 'duration_secs': 0.575018} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.090792] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1057.091115] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee82b88-04fa-4cf9-9c41-a42a17af465c tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance 'ada96855-b35f-4ff5-b3ce-e4fa1baaedc4' progress to 100 {{(pid=61964) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.165792] env[61964]: DEBUG nova.scheduler.client.report [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1057.273720] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.273934] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.274105] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1057.669949] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.670541] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1057.673316] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.053s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.673605] env[61964]: DEBUG nova.objects.instance [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lazy-loading 'resources' on Instance uuid ee273d32-a49f-43b3-8b08-8017eb1b1636 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.686801] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1057.687075] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35e72770-227e-400e-ba61-79e58a5ef5d0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.693949] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1057.693949] env[61964]: value = "task-1041564" [ 1057.693949] env[61964]: _type = "Task" [ 1057.693949] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.702329] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.178262] env[61964]: DEBUG nova.compute.utils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.184030] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1058.184254] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1058.206397] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041564, 'name': PowerOffVM_Task, 'duration_secs': 0.44292} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.209730] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.210515] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1058.210763] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.210942] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1058.211180] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.211345] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1058.211503] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1058.212482] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1058.212482] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1058.212482] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1058.212985] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1058.212985] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1058.219177] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c9c9713-6ee6-440c-bedf-7856479e314c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.237117] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1058.237117] env[61964]: value = "task-1041565" [ 1058.237117] env[61964]: _type = "Task" [ 1058.237117] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.246210] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041565, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.276472] env[61964]: DEBUG nova.policy [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf85b27f46c94956982ca3c713aef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b79a033963e04e02a98df18f8429263f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1058.360466] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd95588-3f39-4ba9-97b1-710f2b907891 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.368383] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80bb422-3498-4524-892b-7d6c96d0d4d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.400099] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c48530b-c3e6-4868-93f5-4e709131a5f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.407742] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edcc672-8597-4d0d-8632-01ae1dd6eaa2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.421162] env[61964]: DEBUG nova.compute.provider_tree [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.647301] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Successfully created port: e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1058.684728] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1058.751252] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041565, 'name': ReconfigVM_Task, 'duration_secs': 0.329855} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.752294] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9983eb2a-757f-42f8-b625-a68704cbbd1d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.777709] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1058.778051] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.778264] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1058.778463] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.778735] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1058.778801] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1058.779055] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1058.779260] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1058.779444] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1058.779649] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1058.779837] env[61964]: DEBUG nova.virt.hardware [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1058.780723] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ff5cf13-89b4-4a57-9d99-c66a3c0fe3f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.788805] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1058.788805] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5275e934-ba3b-c0b8-9f5b-4d475c4f100b" [ 1058.788805] env[61964]: _type = "Task" [ 1058.788805] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.799346] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5275e934-ba3b-c0b8-9f5b-4d475c4f100b, 'name': SearchDatastore_Task, 'duration_secs': 0.007222} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.805504] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1058.806146] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-253b0a32-289f-401e-9d52-b61565fe9f41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.828414] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1058.828414] env[61964]: value = "task-1041566" [ 1058.828414] env[61964]: _type = "Task" [ 1058.828414] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.841637] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041566, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.924473] env[61964]: DEBUG nova.scheduler.client.report [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.340654] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041566, 'name': ReconfigVM_Task, 'duration_secs': 0.248503} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.341331] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1059.342968] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc2881a-23b0-4307-846b-d9b175862f10 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.370752] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.371164] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e915132-d89d-43da-b17f-ed41ce3e45d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.392854] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1059.392854] env[61964]: value = "task-1041567" [ 1059.392854] env[61964]: _type = "Task" [ 1059.392854] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.401544] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.429763] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.432429] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.405s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.432671] env[61964]: DEBUG nova.objects.instance [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lazy-loading 'resources' on Instance uuid 204fc3b1-c4a6-43ef-a408-e41be883034e {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.454443] env[61964]: INFO nova.scheduler.client.report [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Deleted allocations for instance ee273d32-a49f-43b3-8b08-8017eb1b1636 [ 1059.695441] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1059.708072] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1059.708420] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230572', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'name': 'volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'serial': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1059.709619] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a59daa6-8f5a-4355-b3b3-e50e615c2120 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.729132] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1059.729420] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.729651] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1059.729891] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.730129] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1059.730333] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1059.730612] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1059.730834] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1059.731110] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1059.731346] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1059.731940] env[61964]: DEBUG nova.virt.hardware [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1059.732488] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4701c4e3-dc35-4fdb-9557-c134de3a3e0d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.735521] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e78eff4-aa76-4a77-bffa-58ddd2c7dc89 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.764410] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff/volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.766946] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbc02a75-d1fc-403d-8f50-6ffb57b976a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.781011] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cd93e9-c9e5-4512-be1b-0d003a605eb7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.798567] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1059.798567] env[61964]: value = "task-1041568" [ 1059.798567] env[61964]: _type = "Task" [ 1059.798567] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.807484] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041568, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.903172] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041567, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.961814] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a85ef6cb-7dbe-42ae-a6f6-2aeb152b1f78 tempest-InstanceActionsTestJSON-2114180305 tempest-InstanceActionsTestJSON-2114180305-project-member] Lock "ee273d32-a49f-43b3-8b08-8017eb1b1636" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.731s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.996090] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.996381] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.996582] env[61964]: DEBUG nova.compute.manager [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Going to confirm migration 6 {{(pid=61964) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5218}} [ 1060.068099] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4538e8f-8f13-407a-a699-2b2a3698a85e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.075982] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa4b0ff-e642-499c-9f26-c24646d960cb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.109044] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5450499c-d8f2-4142-a5f4-326aed9ef970 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.116657] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86170fa1-98b9-4956-8a62-2d8e9dc594c7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.133022] env[61964]: DEBUG nova.compute.provider_tree [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.312033] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041568, 'name': ReconfigVM_Task, 'duration_secs': 0.408464} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.312033] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfigured VM instance instance-00000061 to attach disk [datastore1] volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff/volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.317722] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0033ce0-5c13-44e1-ac84-29d734bfdb18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.333105] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1060.333105] env[61964]: value = "task-1041569" [ 1060.333105] env[61964]: _type = "Task" [ 1060.333105] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.341343] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041569, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.345704] env[61964]: DEBUG nova.compute.manager [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Received event network-vif-plugged-e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1060.345952] env[61964]: DEBUG oslo_concurrency.lockutils [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.346138] env[61964]: DEBUG oslo_concurrency.lockutils [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] Lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.346334] env[61964]: DEBUG oslo_concurrency.lockutils [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] Lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.346511] env[61964]: DEBUG nova.compute.manager [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] No waiting events found dispatching network-vif-plugged-e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1060.346691] env[61964]: WARNING nova.compute.manager [req-0b46a333-7026-4705-b9c2-c12db5cc4f5b req-9cc772d6-0990-43cf-822a-087c769c9d78 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Received unexpected event network-vif-plugged-e5ee90c2-06be-418d-b67e-9257fb728e47 for instance with vm_state building and task_state spawning. [ 1060.403508] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041567, 'name': ReconfigVM_Task, 'duration_secs': 0.551805} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.403801] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186/15e92512-3c92-43f0-891c-84bdaad1d186.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.404779] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc13e23-9b76-4236-a0d9-e2ee43081925 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.429023] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89dc8e1b-256f-4bbc-999b-428995e38d58 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.453080] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Successfully updated port: e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1060.456099] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f387e957-aeac-43b9-b50c-bb0b3043c315 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.481976] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046e681a-9956-4aab-b2c5-26b7c107b282 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.491444] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1060.492250] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f3f33b7-aa55-4789-bf3d-96bbf6601f3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.498816] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1060.498816] env[61964]: value = "task-1041570" [ 1060.498816] env[61964]: _type = "Task" [ 1060.498816] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.513965] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.624752] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.624992] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquired lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.625272] env[61964]: DEBUG nova.network.neutron [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.625589] env[61964]: DEBUG nova.objects.instance [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'info_cache' on Instance uuid ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.637211] env[61964]: DEBUG nova.scheduler.client.report [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1060.807132] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Didn't find any instances for network info cache update. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10456}} [ 1060.807364] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.807526] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.807674] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.807868] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.808048] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.808203] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.808336] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1060.808479] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1060.843091] env[61964]: DEBUG oslo_vmware.api [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041569, 'name': ReconfigVM_Task, 'duration_secs': 0.139827} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.843637] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230572', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'name': 'volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'serial': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1060.955968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.956138] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.956296] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1061.009828] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.144110] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.173206] env[61964]: INFO nova.scheduler.client.report [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted allocations for instance 204fc3b1-c4a6-43ef-a408-e41be883034e [ 1061.311554] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.312486] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.312711] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.312947] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1061.313941] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a53e3a-0e0d-4286-acdd-30207c90d3b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.326857] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8aa47e-2692-4cdd-959a-82c14d763d97 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.342308] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce19e030-1f34-48c6-98ca-b14e32daf651 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.352176] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807f5466-fb11-4141-8586-187788ffc5a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.383600] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180586MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1061.383767] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.383977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.504539] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1061.512013] env[61964]: DEBUG oslo_vmware.api [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041570, 'name': PowerOnVM_Task, 'duration_secs': 0.775639} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.512013] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1061.686095] env[61964]: DEBUG oslo_concurrency.lockutils [None req-007076dc-644a-4741-a42b-2604c02af6a8 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "204fc3b1-c4a6-43ef-a408-e41be883034e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.048s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.834353] env[61964]: DEBUG nova.network.neutron [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [{"id": "139c3706-9f15-4801-af8d-879960814bd9", "address": "fa:16:3e:da:a2:2f", "network": {"id": "10e45ff1-99f7-4149-b379-1722cfe026b9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1157079028-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb92e5b578d54f1499b00aa08e7841c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap139c3706-9f", "ovs_interfaceid": "139c3706-9f15-4801-af8d-879960814bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.894034] env[61964]: DEBUG nova.objects.instance [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.926178] env[61964]: DEBUG nova.network.neutron [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating instance_info_cache with network_info: [{"id": "e5ee90c2-06be-418d-b67e-9257fb728e47", "address": "fa:16:3e:ab:78:f7", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5ee90c2-06", "ovs_interfaceid": "e5ee90c2-06be-418d-b67e-9257fb728e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.338417] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Releasing lock "refresh_cache-ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.338676] env[61964]: DEBUG nova.objects.instance [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lazy-loading 'migration_context' on Instance uuid ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.401060] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Applying migration context for instance ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 as it has an incoming, in-progress migration 90eac758-d932-4603-9dd4-78a03c6fccac. Migration status is finished {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1062.402195] env[61964]: INFO nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating resource usage from migration 90eac758-d932-4603-9dd4-78a03c6fccac [ 1062.407294] env[61964]: DEBUG oslo_concurrency.lockutils [None req-b1ccf583-3668-4115-9302-f310791497b1 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.300s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.425177] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 0da26d16-84f2-4c00-97f5-ff132fa122d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.425415] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 2da49b28-0cb0-4486-8e69-ce6fba20387c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.425577] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 15e92512-3c92-43f0-891c-84bdaad1d186 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.425730] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 6ad71a14-c3d3-4db3-968a-d51d51d86496 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.425879] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 9b48b942-5e29-4fea-83b7-a1551f78345c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.426019] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Migration 90eac758-d932-4603-9dd4-78a03c6fccac is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1062.426140] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.426257] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 738ae769-4ebd-4828-9414-1ba8309833d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.426496] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1062.426634] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1062.429484] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.429772] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Instance network_info: |[{"id": "e5ee90c2-06be-418d-b67e-9257fb728e47", "address": "fa:16:3e:ab:78:f7", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5ee90c2-06", "ovs_interfaceid": "e5ee90c2-06be-418d-b67e-9257fb728e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1062.430433] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:78:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97b68ed7-8461-4345-b064-96a1dde53a86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5ee90c2-06be-418d-b67e-9257fb728e47', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.438193] env[61964]: DEBUG oslo.service.loopingcall [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.438703] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1062.439103] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc79ad17-fc24-49c9-be08-0fb12ae11223 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.469530] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.469530] env[61964]: value = "task-1041571" [ 1062.469530] env[61964]: _type = "Task" [ 1062.469530] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.482317] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041571, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.544688] env[61964]: DEBUG nova.compute.manager [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Received event network-changed-e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1062.544951] env[61964]: DEBUG nova.compute.manager [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Refreshing instance network info cache due to event network-changed-e5ee90c2-06be-418d-b67e-9257fb728e47. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1062.545189] env[61964]: DEBUG oslo_concurrency.lockutils [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] Acquiring lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.545360] env[61964]: DEBUG oslo_concurrency.lockutils [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] Acquired lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.545496] env[61964]: DEBUG nova.network.neutron [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Refreshing network info cache for port e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.559473] env[61964]: INFO nova.compute.manager [None req-29eb8111-fc87-4883-b048-fb69beee718e tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance to original state: 'active' [ 1062.564942] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612b6083-d6ab-45b4-9236-69cfb5605bbb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.573287] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02b5119-8bc8-45cc-b0d3-20c1eef09f7c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.606554] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff77eb3-517d-4a46-8cee-61e22563321b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.614686] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e2af86-3448-468f-94c0-47222bc7d8c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.628827] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.840607] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.840887] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.841108] env[61964]: INFO nova.compute.manager [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Shelving [ 1062.842891] env[61964]: DEBUG nova.objects.base [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1062.843986] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95f71bb-6d26-45d6-bd0f-e5f9792fc383 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.863576] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1083827b-4bb3-4d75-9856-2985016d479d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.869494] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1062.869494] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5267a6a6-0659-070a-ea83-d6becbcfa971" [ 1062.869494] env[61964]: _type = "Task" [ 1062.869494] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.879897] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5267a6a6-0659-070a-ea83-d6becbcfa971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.979361] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041571, 'name': CreateVM_Task} progress is 99%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.131965] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1063.352903] env[61964]: DEBUG nova.network.neutron [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updated VIF entry in instance network info cache for port e5ee90c2-06be-418d-b67e-9257fb728e47. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.353339] env[61964]: DEBUG nova.network.neutron [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating instance_info_cache with network_info: [{"id": "e5ee90c2-06be-418d-b67e-9257fb728e47", "address": "fa:16:3e:ab:78:f7", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5ee90c2-06", "ovs_interfaceid": "e5ee90c2-06be-418d-b67e-9257fb728e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.380729] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5267a6a6-0659-070a-ea83-d6becbcfa971, 'name': SearchDatastore_Task, 'duration_secs': 0.010079} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.381063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.480465] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041571, 'name': CreateVM_Task, 'duration_secs': 0.556709} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.480839] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1063.481346] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.481522] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.481850] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.482151] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f332a3d8-8a1b-4d9b-8693-b006c0baf65c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.486715] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1063.486715] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ffa23-3b9c-01a0-f641-cac33840cfd4" [ 1063.486715] env[61964]: _type = "Task" [ 1063.486715] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.494598] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ffa23-3b9c-01a0-f641-cac33840cfd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.523586] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.523910] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.637022] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1063.637273] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.253s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.637568] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.257s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.851933] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1063.851933] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cba9653-58d5-4467-93f8-9901001c8a8f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.855501] env[61964]: DEBUG oslo_concurrency.lockutils [req-04ffe424-0cd5-4cd4-ac6a-d79ee70a602b req-27dcb8a6-5b3c-4a19-93fe-1381bd8e85a6 service nova] Releasing lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.859227] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1063.859227] env[61964]: value = "task-1041572" [ 1063.859227] env[61964]: _type = "Task" [ 1063.859227] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.867020] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.927537] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.927889] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.928163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.928365] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.928545] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.931128] env[61964]: INFO nova.compute.manager [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Terminating instance [ 1063.996874] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]523ffa23-3b9c-01a0-f641-cac33840cfd4, 'name': SearchDatastore_Task, 'duration_secs': 0.00919} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.997532] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.997532] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.997757] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.997972] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.998141] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.998441] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e20ec08e-7624-4f20-b80c-e3e0b200ac8d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.006557] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.006803] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1064.008217] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3118c381-ff8a-462f-80de-0b7a435f9157 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.013600] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1064.013600] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293736c-2eaf-9619-5e77-1cb2e5ac8ffa" [ 1064.013600] env[61964]: _type = "Task" [ 1064.013600] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.021660] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293736c-2eaf-9619-5e77-1cb2e5ac8ffa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.026509] env[61964]: INFO nova.compute.manager [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Detaching volume 63d77f81-da29-4cef-a84c-71585e99374b [ 1064.068238] env[61964]: INFO nova.virt.block_device [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attempting to driver detach volume 63d77f81-da29-4cef-a84c-71585e99374b from mountpoint /dev/sdb [ 1064.068620] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1064.068855] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230570', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'name': 'volume-63d77f81-da29-4cef-a84c-71585e99374b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'serial': '63d77f81-da29-4cef-a84c-71585e99374b'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1064.069889] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e32c859-6f70-482c-b8b7-ac2c312de8a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.094769] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e087c3b-9388-4902-ae8d-8f4a1d840f4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.104906] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6242d864-e779-482c-bb42-4a3cda019e4c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.130380] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edbde17-13c9-4f71-b6f9-eea4c89c6a30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.150324] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] The volume has not been displaced from its original location: [datastore2] volume-63d77f81-da29-4cef-a84c-71585e99374b/volume-63d77f81-da29-4cef-a84c-71585e99374b.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1064.155770] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1064.157456] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6127b5f-0d0d-452f-8e21-0b4f8d044f30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.176418] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1064.176418] env[61964]: value = "task-1041573" [ 1064.176418] env[61964]: _type = "Task" [ 1064.176418] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.190927] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041573, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.277446] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2aece8-5a50-4050-b367-e2b029e14db3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.285521] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6417f9-c732-44a8-948b-ad6119a437de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.315592] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a3f36f-932a-48c0-bea3-ca2892c1d4bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.323072] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2424642-a527-4088-94bc-055ba82c0173 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.336177] env[61964]: DEBUG nova.compute.provider_tree [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.368710] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041572, 'name': PowerOffVM_Task, 'duration_secs': 0.200317} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.368960] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1064.369740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efae08b-406d-4921-bc59-1a34c9f15c9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.388883] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa96c9cf-d8f3-4659-b6c0-9df4fe95295a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.435155] env[61964]: DEBUG nova.compute.manager [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1064.435852] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1064.436187] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c111dae0-fca6-4357-90e7-bccb903eb14b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.442422] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1064.442422] env[61964]: value = "task-1041574" [ 1064.442422] env[61964]: _type = "Task" [ 1064.442422] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.451950] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.523870] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5293736c-2eaf-9619-5e77-1cb2e5ac8ffa, 'name': SearchDatastore_Task, 'duration_secs': 0.021394} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.524835] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42325909-c48c-4b8e-8381-1ff55054979b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.530496] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1064.530496] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520ba0af-66a2-cea5-c62b-0bed495ec19d" [ 1064.530496] env[61964]: _type = "Task" [ 1064.530496] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.539497] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.539815] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.540057] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.540257] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.540431] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.542084] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520ba0af-66a2-cea5-c62b-0bed495ec19d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.542809] env[61964]: INFO nova.compute.manager [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Terminating instance [ 1064.604869] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "9b48b942-5e29-4fea-83b7-a1551f78345c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.605165] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.605381] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.605630] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.605833] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.608282] env[61964]: INFO nova.compute.manager [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Terminating instance [ 1064.690443] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041573, 'name': ReconfigVM_Task, 'duration_secs': 0.231757} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.690733] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1064.695480] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62a1d24c-2c51-4d2c-bdaf-df3b77c0c268 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.709718] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1064.709718] env[61964]: value = "task-1041575" [ 1064.709718] env[61964]: _type = "Task" [ 1064.709718] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.717325] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041575, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.839230] env[61964]: DEBUG nova.scheduler.client.report [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.898950] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1064.899235] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-777b2a67-e4e8-4320-865c-4c9396749ae8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.907238] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1064.907238] env[61964]: value = "task-1041576" [ 1064.907238] env[61964]: _type = "Task" [ 1064.907238] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.914754] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041576, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.952244] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041574, 'name': PowerOffVM_Task, 'duration_secs': 0.25371} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.952526] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1064.952732] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1064.952931] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230557', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'name': 'volume-b92af457-aeb9-4914-a6b6-8291bbfc0491', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '15e92512-3c92-43f0-891c-84bdaad1d186', 'attached_at': '2024-09-30T16:29:05.000000', 'detached_at': '', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'serial': 'b92af457-aeb9-4914-a6b6-8291bbfc0491'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1064.953718] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b96a87-3a23-475e-a6a5-3e4df97f2b5c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.973760] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c864786a-fb96-4e90-8005-c5da60fe55bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.981780] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f6d09f-26da-4462-b08d-3b48f79d1dd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.002893] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e5fdf0-e4c3-464d-8a00-0bb96139a456 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.017811] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] The volume has not been displaced from its original location: [datastore1] volume-b92af457-aeb9-4914-a6b6-8291bbfc0491/volume-b92af457-aeb9-4914-a6b6-8291bbfc0491.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1065.022990] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1065.023309] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da4c33c0-6c17-4a4e-8899-d2dd0c82a5c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.044288] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]520ba0af-66a2-cea5-c62b-0bed495ec19d, 'name': SearchDatastore_Task, 'duration_secs': 0.0098} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.045495] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.045799] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 738ae769-4ebd-4828-9414-1ba8309833d6/738ae769-4ebd-4828-9414-1ba8309833d6.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1065.046134] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1065.046134] env[61964]: value = "task-1041577" [ 1065.046134] env[61964]: _type = "Task" [ 1065.046134] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.046672] env[61964]: DEBUG nova.compute.manager [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1065.046871] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1065.047112] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dee5125-8fd9-443e-8409-9f6e4ecf07a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.049404] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b273e53-9342-4433-a9b3-b71299172bcf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.060191] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041577, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.063525] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1065.063525] env[61964]: value = "task-1041578" [ 1065.063525] env[61964]: _type = "Task" [ 1065.063525] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.063831] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1065.064058] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-880f7534-cd76-4b0e-84d8-74b56e336553 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.074606] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.075910] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1065.075910] env[61964]: value = "task-1041579" [ 1065.075910] env[61964]: _type = "Task" [ 1065.075910] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.083257] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.113510] env[61964]: DEBUG nova.compute.manager [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1065.113793] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1065.115216] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a5ee04-3da6-4697-a3f8-051fb9d5d4fc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.123263] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1065.123573] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47f2f63c-dd3b-42e5-941b-0815421a83a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.130349] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1065.130349] env[61964]: value = "task-1041580" [ 1065.130349] env[61964]: _type = "Task" [ 1065.130349] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.138690] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.221765] env[61964]: DEBUG oslo_vmware.api [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041575, 'name': ReconfigVM_Task, 'duration_secs': 0.136652} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.222170] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230570', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'name': 'volume-63d77f81-da29-4cef-a84c-71585e99374b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': '63d77f81-da29-4cef-a84c-71585e99374b', 'serial': '63d77f81-da29-4cef-a84c-71585e99374b'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1065.417687] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041576, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.559944] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041577, 'name': ReconfigVM_Task, 'duration_secs': 0.255149} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.560388] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1065.565189] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cd70b8e-466c-44a5-a4ed-5765b3f9d763 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.583811] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1065.583811] env[61964]: value = "task-1041581" [ 1065.583811] env[61964]: _type = "Task" [ 1065.583811] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.589437] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041578, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471449} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.589693] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041579, 'name': PowerOffVM_Task, 'duration_secs': 0.197092} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.592440] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 738ae769-4ebd-4828-9414-1ba8309833d6/738ae769-4ebd-4828-9414-1ba8309833d6.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1065.592669] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.592923] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1065.593111] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1065.593342] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14f92662-a640-4ed5-a46c-026fd04091ed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.595155] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9a070b5-218c-4722-8365-e34f9087cc38 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.600684] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.601788] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1065.601788] env[61964]: value = "task-1041582" [ 1065.601788] env[61964]: _type = "Task" [ 1065.601788] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.610424] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041582, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.639883] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041580, 'name': PowerOffVM_Task, 'duration_secs': 0.216817} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.640243] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1065.640454] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1065.640736] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc7029b4-e468-42b5-a163-a5b0713fc010 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.671197] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1065.671436] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1065.671761] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleting the datastore file [datastore2] 6ad71a14-c3d3-4db3-968a-d51d51d86496 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.672041] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fc15083-8fce-4329-ad97-0fb3a4f02794 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.678285] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1065.678285] env[61964]: value = "task-1041585" [ 1065.678285] env[61964]: _type = "Task" [ 1065.678285] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.686773] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.708252] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1065.708513] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1065.708697] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleting the datastore file [datastore2] 9b48b942-5e29-4fea-83b7-a1551f78345c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.709017] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd30bb14-e4ab-4238-86a3-a51cec9c8269 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.715018] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for the task: (returnval){ [ 1065.715018] env[61964]: value = "task-1041586" [ 1065.715018] env[61964]: _type = "Task" [ 1065.715018] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.722923] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.768213] env[61964]: DEBUG nova.objects.instance [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.850749] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.213s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.917753] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041576, 'name': CreateSnapshot_Task, 'duration_secs': 0.872727} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.918051] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1065.919017] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d0e5d7-8f03-4ad3-9204-a6c330e8a86c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.098584] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041581, 'name': ReconfigVM_Task, 'duration_secs': 0.150637} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.098958] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230557', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'name': 'volume-b92af457-aeb9-4914-a6b6-8291bbfc0491', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '15e92512-3c92-43f0-891c-84bdaad1d186', 'attached_at': '2024-09-30T16:29:05.000000', 'detached_at': '', 'volume_id': 'b92af457-aeb9-4914-a6b6-8291bbfc0491', 'serial': 'b92af457-aeb9-4914-a6b6-8291bbfc0491'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1066.099291] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1066.100060] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e1a4ee-c3e8-4d95-b0ca-325095c09fba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.109282] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1066.111862] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03d8cf67-973f-4432-a589-939254316639 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.113159] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041582, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063307} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.113404] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.114399] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3b2866-66fa-4f13-8388-25be7311a6df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.135759] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 738ae769-4ebd-4828-9414-1ba8309833d6/738ae769-4ebd-4828-9414-1ba8309833d6.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.136053] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efd96afb-5825-4bbe-869a-990bda4b6398 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.154313] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1066.154313] env[61964]: value = "task-1041588" [ 1066.154313] env[61964]: _type = "Task" [ 1066.154313] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.165134] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.188242] env[61964]: DEBUG oslo_vmware.api [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176859} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.189555] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.189839] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1066.190048] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1066.190347] env[61964]: INFO nova.compute.manager [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1066.190562] env[61964]: DEBUG oslo.service.loopingcall [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.190823] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1066.191043] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1066.191251] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleting the datastore file [datastore1] 15e92512-3c92-43f0-891c-84bdaad1d186 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1066.191520] env[61964]: DEBUG nova.compute.manager [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1066.191636] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1066.193557] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3d972f4-b375-45ca-89f1-5f27d1e550b7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.200471] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1066.200471] env[61964]: value = "task-1041589" [ 1066.200471] env[61964]: _type = "Task" [ 1066.200471] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.208369] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.224221] env[61964]: DEBUG oslo_vmware.api [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Task: {'id': task-1041586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178209} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.226630] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.226849] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1066.227068] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1066.227270] env[61964]: INFO nova.compute.manager [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1066.227518] env[61964]: DEBUG oslo.service.loopingcall [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.227785] env[61964]: DEBUG nova.compute.manager [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1066.227884] env[61964]: DEBUG nova.network.neutron [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1066.443563] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1066.445373] env[61964]: INFO nova.scheduler.client.report [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocation for migration 90eac758-d932-4603-9dd4-78a03c6fccac [ 1066.450817] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-749bece0-695b-4615-bfe9-0ba402d8d7f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.459386] env[61964]: DEBUG nova.compute.manager [req-0b2d8518-c723-432f-a8aa-8c766372737f req-4a0c6e0a-3804-44d9-a9f3-e38ab348d793 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Received event network-vif-deleted-9afcca07-e043-46d7-b521-2fa0062b8e87 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1066.459612] env[61964]: INFO nova.compute.manager [req-0b2d8518-c723-432f-a8aa-8c766372737f req-4a0c6e0a-3804-44d9-a9f3-e38ab348d793 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Neutron deleted interface 9afcca07-e043-46d7-b521-2fa0062b8e87; detaching it from the instance and deleting it from the info cache [ 1066.459825] env[61964]: DEBUG nova.network.neutron [req-0b2d8518-c723-432f-a8aa-8c766372737f req-4a0c6e0a-3804-44d9-a9f3-e38ab348d793 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.462618] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1066.462618] env[61964]: value = "task-1041590" [ 1066.462618] env[61964]: _type = "Task" [ 1066.462618] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.474248] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.477905] env[61964]: DEBUG nova.compute.manager [req-e761052a-e9d3-4b67-9721-ba90df2df7c1 req-290100f2-f352-439e-9fe7-8df987dcf3ec service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Received event network-vif-deleted-f0603f94-32f6-46df-a2fe-498227b76538 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1066.478089] env[61964]: INFO nova.compute.manager [req-e761052a-e9d3-4b67-9721-ba90df2df7c1 req-290100f2-f352-439e-9fe7-8df987dcf3ec service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Neutron deleted interface f0603f94-32f6-46df-a2fe-498227b76538; detaching it from the instance and deleting it from the info cache [ 1066.478269] env[61964]: DEBUG nova.network.neutron [req-e761052a-e9d3-4b67-9721-ba90df2df7c1 req-290100f2-f352-439e-9fe7-8df987dcf3ec service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.664395] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041588, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.709955] env[61964]: DEBUG oslo_vmware.api [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166552} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.710235] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.710428] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1066.710612] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1066.710796] env[61964]: INFO nova.compute.manager [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1066.711055] env[61964]: DEBUG oslo.service.loopingcall [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.711253] env[61964]: DEBUG nova.compute.manager [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1066.711350] env[61964]: DEBUG nova.network.neutron [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1066.779369] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.782060] env[61964]: DEBUG oslo_concurrency.lockutils [None req-aab0a64c-ba4a-44cf-8c58-f74e52670f13 tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.258s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.783291] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.004s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.921242] env[61964]: DEBUG nova.network.neutron [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.956080] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.960s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.962882] env[61964]: DEBUG nova.network.neutron [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.964627] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9cf2d9e-73e8-436a-a787-e6e0f1e11a6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.980130] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.981484] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84cbb2ba-dba3-4235-bb30-bd415822cf3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.985721] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bf510b-9b84-4562-8aa7-8fee05fc7ed2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.002920] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e8e1d5-6d80-4b20-9377-ce55cd797114 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.020981] env[61964]: DEBUG nova.compute.manager [req-0b2d8518-c723-432f-a8aa-8c766372737f req-4a0c6e0a-3804-44d9-a9f3-e38ab348d793 service nova] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Detach interface failed, port_id=9afcca07-e043-46d7-b521-2fa0062b8e87, reason: Instance 6ad71a14-c3d3-4db3-968a-d51d51d86496 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1067.036137] env[61964]: DEBUG nova.compute.manager [req-e761052a-e9d3-4b67-9721-ba90df2df7c1 req-290100f2-f352-439e-9fe7-8df987dcf3ec service nova] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Detach interface failed, port_id=f0603f94-32f6-46df-a2fe-498227b76538, reason: Instance 9b48b942-5e29-4fea-83b7-a1551f78345c could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1067.166857] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041588, 'name': ReconfigVM_Task, 'duration_secs': 0.791732} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.166857] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 738ae769-4ebd-4828-9414-1ba8309833d6/738ae769-4ebd-4828-9414-1ba8309833d6.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.168228] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cede28d-dd00-43f1-aa34-2368a293f315 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.174102] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1067.174102] env[61964]: value = "task-1041591" [ 1067.174102] env[61964]: _type = "Task" [ 1067.174102] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.183662] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041591, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.289319] env[61964]: INFO nova.compute.manager [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Detaching volume c2808fb3-02c8-4d39-9906-9d4b44c139ff [ 1067.327597] env[61964]: INFO nova.virt.block_device [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Attempting to driver detach volume c2808fb3-02c8-4d39-9906-9d4b44c139ff from mountpoint /dev/sdc [ 1067.327839] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1067.328047] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230572', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'name': 'volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'serial': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1067.328948] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3025224b-d9f6-4be5-9952-033d73dac874 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.354146] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82a7f4f-bdac-48c2-bbeb-9996a3120097 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.361898] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f10dd03-1ac1-4a65-86eb-50cf429b19e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.382240] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2bf6df-0b18-45b6-a7fc-e169054ca685 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.397857] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] The volume has not been displaced from its original location: [datastore1] volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff/volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1067.403294] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfiguring VM instance instance-00000061 to detach disk 2002 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1067.403611] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bef5f63-2eb2-4cb9-813d-f84fb8f093b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.421989] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1067.421989] env[61964]: value = "task-1041592" [ 1067.421989] env[61964]: _type = "Task" [ 1067.421989] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.422469] env[61964]: INFO nova.compute.manager [-] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Took 1.23 seconds to deallocate network for instance. [ 1067.433586] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041592, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.465887] env[61964]: INFO nova.compute.manager [-] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Took 1.24 seconds to deallocate network for instance. [ 1067.479704] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.642127] env[61964]: DEBUG nova.network.neutron [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.684023] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041591, 'name': Rename_Task, 'duration_secs': 0.132183} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.684023] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1067.684340] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c868a85-408b-4aed-8e79-b4429676b9d9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.689442] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1067.689442] env[61964]: value = "task-1041593" [ 1067.689442] env[61964]: _type = "Task" [ 1067.689442] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.697993] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.932491] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.932774] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.933016] env[61964]: DEBUG nova.objects.instance [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lazy-loading 'resources' on Instance uuid 6ad71a14-c3d3-4db3-968a-d51d51d86496 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.934045] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041592, 'name': ReconfigVM_Task, 'duration_secs': 0.259299} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.934493] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Reconfigured VM instance instance-00000061 to detach disk 2002 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.939102] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ac2d404-df0a-4331-b36f-36f90f980444 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.954595] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1067.954595] env[61964]: value = "task-1041594" [ 1067.954595] env[61964]: _type = "Task" [ 1067.954595] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.964772] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041594, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.974825] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.975086] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.025584] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.025843] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.026126] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.026309] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.026466] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.028852] env[61964]: INFO nova.compute.manager [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Terminating instance [ 1068.144910] env[61964]: INFO nova.compute.manager [-] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Took 1.43 seconds to deallocate network for instance. [ 1068.201918] env[61964]: DEBUG oslo_vmware.api [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041593, 'name': PowerOnVM_Task, 'duration_secs': 0.48061} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.202269] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1068.202511] env[61964]: INFO nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1068.202731] env[61964]: DEBUG nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1068.203562] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23931117-e322-4a5c-8074-7dc7e77ec5fa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.465022] env[61964]: DEBUG oslo_vmware.api [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041594, 'name': ReconfigVM_Task, 'duration_secs': 0.125423} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.467438] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230572', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'name': 'volume-c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0da26d16-84f2-4c00-97f5-ff132fa122d2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff', 'serial': 'c2808fb3-02c8-4d39-9906-9d4b44c139ff'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1068.477220] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.504317] env[61964]: DEBUG nova.compute.manager [req-d232b93d-9a0b-40bc-a2fc-edf29f9a8d4a req-dad095e7-74c5-4f94-a8fc-b2343d6a80cf service nova] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Received event network-vif-deleted-da9bd08d-0a86-4d20-82b3-0160a3cf26b2 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1068.530316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7630df70-95d1-42a1-a484-41be08eb2aef {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.533709] env[61964]: DEBUG nova.compute.manager [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1068.533918] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1068.534667] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3469696d-d0ee-4d11-993c-4f2cf9c1ed26 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.542988] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cf9fe9-4e24-44ce-9bff-1db6e8dac0df {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.545850] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1068.546077] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aba5c059-5004-42c6-b7fc-84ec9b1421a2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.577095] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2360774-a745-47ef-92ce-48c97d9ad3f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.579869] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1068.579869] env[61964]: value = "task-1041595" [ 1068.579869] env[61964]: _type = "Task" [ 1068.579869] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.586533] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6a6fb1-22f8-4f61-9582-8676bea32fd2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.592858] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.602086] env[61964]: DEBUG nova.compute.provider_tree [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.695058] env[61964]: INFO nova.compute.manager [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Took 0.55 seconds to detach 1 volumes for instance. [ 1068.721661] env[61964]: INFO nova.compute.manager [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Took 18.42 seconds to build instance. [ 1068.980324] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.008761] env[61964]: DEBUG nova.objects.instance [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'flavor' on Instance uuid 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.089884] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041595, 'name': PowerOffVM_Task, 'duration_secs': 0.371172} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.090173] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1069.090349] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1069.090589] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69f7a12b-41b4-441c-8025-5ad9835081f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.105053] env[61964]: DEBUG nova.scheduler.client.report [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1069.150082] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1069.150418] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1069.150676] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleting the datastore file [datastore2] ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.150993] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e72b406a-8905-420c-b932-b597a1d74195 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.157831] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for the task: (returnval){ [ 1069.157831] env[61964]: value = "task-1041597" [ 1069.157831] env[61964]: _type = "Task" [ 1069.157831] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.166349] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.203510] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.224235] env[61964]: DEBUG oslo_concurrency.lockutils [None req-910cfa44-f2d3-44a1-b3ac-04164ca58a14 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.934s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.480319] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.610455] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.612879] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.638s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.613138] env[61964]: DEBUG nova.objects.instance [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lazy-loading 'resources' on Instance uuid 9b48b942-5e29-4fea-83b7-a1551f78345c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.637876] env[61964]: INFO nova.scheduler.client.report [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted allocations for instance 6ad71a14-c3d3-4db3-968a-d51d51d86496 [ 1069.669135] env[61964]: DEBUG oslo_vmware.api [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Task: {'id': task-1041597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158363} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.669393] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.669584] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1069.669830] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1069.670012] env[61964]: INFO nova.compute.manager [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1069.670264] env[61964]: DEBUG oslo.service.loopingcall [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.670494] env[61964]: DEBUG nova.compute.manager [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1069.670591] env[61964]: DEBUG nova.network.neutron [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1069.980842] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.015885] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6142f638-66a9-42e9-9faf-1a60b118909f tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.232s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.147724] env[61964]: DEBUG oslo_concurrency.lockutils [None req-0be0ecac-f3e1-4195-a7f2-1b7deafca39e tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "6ad71a14-c3d3-4db3-968a-d51d51d86496" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.608s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.216727] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa31ffe9-aa7a-448b-a02c-cbb45f825f5c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.224212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248bf9da-f2b4-4d2f-b90b-b7fe3fcc2f3c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.253954] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c08664-869f-4e5a-b734-6208146779e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.261015] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a020552b-fe2c-4675-b514-6027e5fba68f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.273852] env[61964]: DEBUG nova.compute.provider_tree [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.481292] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.532123] env[61964]: DEBUG nova.compute.manager [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Received event network-changed-e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1070.532367] env[61964]: DEBUG nova.compute.manager [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Refreshing instance network info cache due to event network-changed-e5ee90c2-06be-418d-b67e-9257fb728e47. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1070.532595] env[61964]: DEBUG oslo_concurrency.lockutils [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] Acquiring lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.532742] env[61964]: DEBUG oslo_concurrency.lockutils [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] Acquired lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.532907] env[61964]: DEBUG nova.network.neutron [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Refreshing network info cache for port e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1070.662235] env[61964]: DEBUG nova.network.neutron [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.776797] env[61964]: DEBUG nova.scheduler.client.report [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1070.982110] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.164744] env[61964]: INFO nova.compute.manager [-] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Took 1.49 seconds to deallocate network for instance. [ 1071.187549] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.187860] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.188090] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.188284] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.188466] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.191086] env[61964]: INFO nova.compute.manager [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Terminating instance [ 1071.262272] env[61964]: DEBUG nova.network.neutron [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updated VIF entry in instance network info cache for port e5ee90c2-06be-418d-b67e-9257fb728e47. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1071.262661] env[61964]: DEBUG nova.network.neutron [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating instance_info_cache with network_info: [{"id": "e5ee90c2-06be-418d-b67e-9257fb728e47", "address": "fa:16:3e:ab:78:f7", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5ee90c2-06", "ovs_interfaceid": "e5ee90c2-06be-418d-b67e-9257fb728e47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.281760] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.669s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.284501] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.082s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.284501] env[61964]: DEBUG nova.objects.instance [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'resources' on Instance uuid 15e92512-3c92-43f0-891c-84bdaad1d186 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.303449] env[61964]: INFO nova.scheduler.client.report [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Deleted allocations for instance 9b48b942-5e29-4fea-83b7-a1551f78345c [ 1071.482552] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 95%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.672899] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.698752] env[61964]: DEBUG nova.compute.manager [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1071.698980] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1071.699968] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf60353-ce8f-4adb-9051-3a184354d2d8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.707570] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1071.707815] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f16278c-0c70-466d-a409-cd7c11f06024 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.713602] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1071.713602] env[61964]: value = "task-1041598" [ 1071.713602] env[61964]: _type = "Task" [ 1071.713602] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.722095] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.764984] env[61964]: DEBUG oslo_concurrency.lockutils [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] Releasing lock "refresh_cache-738ae769-4ebd-4828-9414-1ba8309833d6" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.765502] env[61964]: DEBUG nova.compute.manager [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Received event network-vif-deleted-139c3706-9f15-4801-af8d-879960814bd9 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1071.765754] env[61964]: INFO nova.compute.manager [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Neutron deleted interface 139c3706-9f15-4801-af8d-879960814bd9; detaching it from the instance and deleting it from the info cache [ 1071.766034] env[61964]: DEBUG nova.network.neutron [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.810776] env[61964]: DEBUG oslo_concurrency.lockutils [None req-edd0f490-d8e2-432c-9ed6-7c5a3df0b5f6 tempest-ListServersNegativeTestJSON-1702402810 tempest-ListServersNegativeTestJSON-1702402810-project-member] Lock "9b48b942-5e29-4fea-83b7-a1551f78345c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.205s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.864948] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7b6b87-66de-47f8-af42-c2b7167508a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.872271] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fc40f3-3444-4405-882f-9e3880867b66 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.904120] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b173f3-f95f-4798-b9aa-3dec6db6c4de {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.911487] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfb515c-d664-4c53-a87d-6c4a7d27795a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.924902] env[61964]: DEBUG nova.compute.provider_tree [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.982853] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.223915] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041598, 'name': PowerOffVM_Task, 'duration_secs': 0.164509} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.224378] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.224637] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1072.224998] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c310a9d-f402-4925-9704-76bc1f573617 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.269555] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-447efea9-99a7-4b83-8c51-ab0bc25532bf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.279263] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42809263-e21a-4b4d-87a6-0d3c10f71401 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.293515] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1072.293749] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1072.293938] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleting the datastore file [datastore2] 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.294218] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50477ba1-0335-4e00-b38b-aff3e9e137d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.300988] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for the task: (returnval){ [ 1072.300988] env[61964]: value = "task-1041600" [ 1072.300988] env[61964]: _type = "Task" [ 1072.300988] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.309467] env[61964]: DEBUG nova.compute.manager [req-60aa4bed-0931-47af-8170-a2bc9b3bb732 req-e35d94ff-cd64-4d5a-8de0-ff37ccba1cd9 service nova] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Detach interface failed, port_id=139c3706-9f15-4801-af8d-879960814bd9, reason: Instance ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1072.314207] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.427703] env[61964]: DEBUG nova.scheduler.client.report [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1072.483217] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.812293] env[61964]: DEBUG oslo_vmware.api [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Task: {'id': task-1041600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138559} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.812790] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.812790] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1072.812915] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1072.813096] env[61964]: INFO nova.compute.manager [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1072.813351] env[61964]: DEBUG oslo.service.loopingcall [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.813549] env[61964]: DEBUG nova.compute.manager [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1072.813770] env[61964]: DEBUG nova.network.neutron [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1072.933522] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.936063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.263s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.936285] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.960627] env[61964]: INFO nova.scheduler.client.report [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted allocations for instance 15e92512-3c92-43f0-891c-84bdaad1d186 [ 1072.968246] env[61964]: INFO nova.scheduler.client.report [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Deleted allocations for instance ada96855-b35f-4ff5-b3ce-e4fa1baaedc4 [ 1072.986358] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.434451] env[61964]: DEBUG nova.compute.manager [req-f4393cc9-784d-4be8-b2b1-d2fedce5cd69 req-305c00e4-d6fc-4d34-942c-1631b5136437 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Received event network-vif-deleted-ed20338e-f092-4780-aede-3f9c39a0b9ea {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1073.434451] env[61964]: INFO nova.compute.manager [req-f4393cc9-784d-4be8-b2b1-d2fedce5cd69 req-305c00e4-d6fc-4d34-942c-1631b5136437 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Neutron deleted interface ed20338e-f092-4780-aede-3f9c39a0b9ea; detaching it from the instance and deleting it from the info cache [ 1073.434451] env[61964]: DEBUG nova.network.neutron [req-f4393cc9-784d-4be8-b2b1-d2fedce5cd69 req-305c00e4-d6fc-4d34-942c-1631b5136437 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.470700] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5d1234c0-233c-4f44-a162-8998c549949b tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "15e92512-3c92-43f0-891c-84bdaad1d186" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.542s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.482588] env[61964]: DEBUG oslo_concurrency.lockutils [None req-33652a0a-8fa9-4d36-a774-4b2ce52ee28a tempest-DeleteServersTestJSON-1908213317 tempest-DeleteServersTestJSON-1908213317-project-member] Lock "ada96855-b35f-4ff5-b3ce-e4fa1baaedc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.457s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.490878] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041590, 'name': CloneVM_Task, 'duration_secs': 6.793012} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.491317] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Created linked-clone VM from snapshot [ 1073.492217] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a23a800-42d6-4c4b-b7e6-c97a7b097c0b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.503328] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Uploading image 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1073.536743] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1073.536743] env[61964]: value = "vm-230575" [ 1073.536743] env[61964]: _type = "VirtualMachine" [ 1073.536743] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1073.537107] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fdf96bb8-4f7f-41bf-861a-e5ffae1e60a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.544481] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease: (returnval){ [ 1073.544481] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284200c-21ee-e89a-9b05-381454abb8d3" [ 1073.544481] env[61964]: _type = "HttpNfcLease" [ 1073.544481] env[61964]: } obtained for exporting VM: (result){ [ 1073.544481] env[61964]: value = "vm-230575" [ 1073.544481] env[61964]: _type = "VirtualMachine" [ 1073.544481] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1073.544795] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the lease: (returnval){ [ 1073.544795] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284200c-21ee-e89a-9b05-381454abb8d3" [ 1073.544795] env[61964]: _type = "HttpNfcLease" [ 1073.544795] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1073.551703] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1073.551703] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284200c-21ee-e89a-9b05-381454abb8d3" [ 1073.551703] env[61964]: _type = "HttpNfcLease" [ 1073.551703] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1073.901378] env[61964]: DEBUG nova.network.neutron [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.937411] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bf6ca57-ea77-4d1f-889e-d2f50c898024 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.948017] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169be3ae-c2d7-4c8a-a5a4-82f46bd33c4a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.974587] env[61964]: DEBUG nova.compute.manager [req-f4393cc9-784d-4be8-b2b1-d2fedce5cd69 req-305c00e4-d6fc-4d34-942c-1631b5136437 service nova] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Detach interface failed, port_id=ed20338e-f092-4780-aede-3f9c39a0b9ea, reason: Instance 0da26d16-84f2-4c00-97f5-ff132fa122d2 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1074.052703] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.052703] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284200c-21ee-e89a-9b05-381454abb8d3" [ 1074.052703] env[61964]: _type = "HttpNfcLease" [ 1074.052703] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1074.053158] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1074.053158] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5284200c-21ee-e89a-9b05-381454abb8d3" [ 1074.053158] env[61964]: _type = "HttpNfcLease" [ 1074.053158] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1074.053834] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c6e421-756a-4c89-ada2-1247987ac63e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.063881] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1074.064083] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1074.200551] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9379466e-760d-4e89-8d1b-68b7f6696a86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.405673] env[61964]: INFO nova.compute.manager [-] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Took 1.59 seconds to deallocate network for instance. [ 1074.495117] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.495391] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.766377] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.767013] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.767248] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Cleaning up deleted instances with incomplete migration {{(pid=61964) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11708}} [ 1074.913244] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.914128] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.914128] env[61964]: DEBUG nova.objects.instance [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lazy-loading 'resources' on Instance uuid 0da26d16-84f2-4c00-97f5-ff132fa122d2 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.997970] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1075.270419] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.489749] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4376243c-2207-4826-95a9-855d333bf0e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.497450] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05870b6f-0ae5-4eda-ada5-0054f3f4066f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.534275] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d304d38c-31c0-4d67-98c0-571ff82afacf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.542925] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bc80bd-649d-4182-aac0-0da0ade5c879 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.549073] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.557466] env[61964]: DEBUG nova.compute.provider_tree [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.061271] env[61964]: DEBUG nova.scheduler.client.report [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.567996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.570437] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.021s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.572058] env[61964]: INFO nova.compute.claims [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.597394] env[61964]: INFO nova.scheduler.client.report [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Deleted allocations for instance 0da26d16-84f2-4c00-97f5-ff132fa122d2 [ 1076.766917] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.767608] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.767788] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.767954] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.768126] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.768289] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.768441] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1077.105096] env[61964]: DEBUG oslo_concurrency.lockutils [None req-1247f818-4d77-4725-91f2-aebfc92c9c0a tempest-AttachVolumeTestJSON-1751285314 tempest-AttachVolumeTestJSON-1751285314-project-member] Lock "0da26d16-84f2-4c00-97f5-ff132fa122d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.917s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.642420] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ace788-05cc-4837-881f-6843b82ad637 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.650828] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8b1426-94bb-417b-a173-a637184a3444 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.683150] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7dba16-525b-4317-b6b3-e8f602446a5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.691811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25990ed-023f-44eb-80bf-6ee040591e1e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.706803] env[61964]: DEBUG nova.compute.provider_tree [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.766714] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.766901] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1077.767034] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 1078.210201] env[61964]: DEBUG nova.scheduler.client.report [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.270963] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Skipping network cache update for instance because it is Building. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10383}} [ 1078.271306] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.272445] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.272445] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1078.272445] env[61964]: DEBUG nova.objects.instance [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lazy-loading 'info_cache' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.716250] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.146s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.716783] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1079.222826] env[61964]: DEBUG nova.compute.utils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.224750] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1079.224965] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.271637] env[61964]: DEBUG nova.policy [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '430778d9f17e4579a81902608fb3501a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b627b25af0c4982a22c8c0bd28641a3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.567324] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Successfully created port: 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.728923] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1080.009953] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.513256] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.513444] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 1080.513775] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.741257] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1080.769212] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1080.769502] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.769675] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1080.769877] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.770061] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1080.770250] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1080.770486] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1080.770663] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1080.770843] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1080.771030] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1080.771230] env[61964]: DEBUG nova.virt.hardware [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1080.772179] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e55788f-27a5-41bb-b141-07014bc90a2e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.781806] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2500f9e-f2b8-426c-9e71-3a4ff7ee3e9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.956669] env[61964]: DEBUG nova.compute.manager [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1080.956669] env[61964]: DEBUG oslo_concurrency.lockutils [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.956669] env[61964]: DEBUG oslo_concurrency.lockutils [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.957085] env[61964]: DEBUG oslo_concurrency.lockutils [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.957085] env[61964]: DEBUG nova.compute.manager [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] No waiting events found dispatching network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1080.957894] env[61964]: WARNING nova.compute.manager [req-3205afdb-7334-466b-9c8e-e8c00cb64e5f req-27a28336-2f6b-40fa-881e-355ab5e1a9c4 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received unexpected event network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 for instance with vm_state building and task_state spawning. [ 1081.019639] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.019868] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.020065] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.020408] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1081.021333] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8113bc90-f6ae-4fc3-8cad-56b158b47b13 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.031027] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257ad8e8-5021-48ef-b315-f5d551495c62 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.044475] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Successfully updated port: 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.046155] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44b64ba-8f39-4565-bb2c-18fa16325764 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.053437] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ede78a-955d-4ace-bc42-4665260afc5f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.083223] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180586MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1081.083498] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.083829] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.550271] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.550415] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.550554] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.888724] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1081.890311] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd390bd-793f-4ca8-a21f-ef5da215b493 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.896150] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1081.896384] env[61964]: ERROR oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk due to incomplete transfer. [ 1081.896603] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2098f848-383b-4b57-999f-50d624b96242 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.903662] env[61964]: DEBUG oslo_vmware.rw_handles [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e4d10a-5559-73cf-ac89-e5ea9cc3334f/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1081.903858] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Uploaded image 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1081.906334] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1081.906528] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-862906fc-d484-43ab-bc8a-2ba32c6266eb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.912431] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1081.912431] env[61964]: value = "task-1041605" [ 1081.912431] env[61964]: _type = "Task" [ 1081.912431] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.919722] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041605, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.084459] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1082.223917] env[61964]: DEBUG nova.network.neutron [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.238748] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 2da49b28-0cb0-4486-8e69-ce6fba20387c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1082.238925] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 738ae769-4ebd-4828-9414-1ba8309833d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1082.239064] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1082.239246] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1082.239385] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1082.283590] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54121c2-f2c4-4a29-b62f-5024ce1baeee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.290721] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1519393-9d87-4215-97af-e6dd68663731 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.320400] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a542d41-b4d3-4171-923b-80aa7ffa82f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.327703] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5223ce3b-387b-41b8-b04e-871ddf85c2e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.340446] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.421740] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041605, 'name': Destroy_Task, 'duration_secs': 0.332359} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.422011] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Destroyed the VM [ 1082.422322] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1082.422572] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-30a33e69-cc4c-4cdf-8c60-1da13a797fe4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.428636] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1082.428636] env[61964]: value = "task-1041606" [ 1082.428636] env[61964]: _type = "Task" [ 1082.428636] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.437009] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041606, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.727300] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.727647] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance network_info: |[{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1082.728174] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:2a:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d7c51d7-90bb-4a31-8ce4-230157e98433', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.735479] env[61964]: DEBUG oslo.service.loopingcall [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.735729] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1082.735992] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cf60aae-2180-4c3b-a7cc-feb4e648183f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.755317] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.755317] env[61964]: value = "task-1041607" [ 1082.755317] env[61964]: _type = "Task" [ 1082.755317] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.766372] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041607, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.843610] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1082.937895] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041606, 'name': RemoveSnapshot_Task, 'duration_secs': 0.354129} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.938265] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1082.938497] env[61964]: DEBUG nova.compute.manager [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1082.939292] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa73adbf-2e8a-49ec-981e-acd791314e8b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.984393] env[61964]: DEBUG nova.compute.manager [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1082.984640] env[61964]: DEBUG nova.compute.manager [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing instance network info cache due to event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1082.984833] env[61964]: DEBUG oslo_concurrency.lockutils [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.985074] env[61964]: DEBUG oslo_concurrency.lockutils [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.985375] env[61964]: DEBUG nova.network.neutron [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.265636] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041607, 'name': CreateVM_Task, 'duration_secs': 0.339978} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.265856] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1083.266573] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.266810] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.267179] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.267434] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ef446d-8173-4232-adaa-cc54aae4cafc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.271638] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1083.271638] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d66500-c205-a052-8b83-5821f12a8195" [ 1083.271638] env[61964]: _type = "Task" [ 1083.271638] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.279082] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d66500-c205-a052-8b83-5821f12a8195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.348920] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1083.349099] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.265s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.349310] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.349448] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Cleaning up deleted instances {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11670}} [ 1083.452072] env[61964]: INFO nova.compute.manager [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Shelve offloading [ 1083.677091] env[61964]: DEBUG nova.network.neutron [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updated VIF entry in instance network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.677511] env[61964]: DEBUG nova.network.neutron [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.781920] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52d66500-c205-a052-8b83-5821f12a8195, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.782258] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.782531] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.782805] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.782968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.783186] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.783483] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b9daf9d-6d54-4761-8989-e6d530db5508 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.791565] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.791741] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1083.792476] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81d0659b-f828-43df-9e8f-1587eb14ca9e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.796958] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1083.796958] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527c9d93-7cd0-f29c-daf7-5561a22a0c1d" [ 1083.796958] env[61964]: _type = "Task" [ 1083.796958] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.804109] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527c9d93-7cd0-f29c-daf7-5561a22a0c1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.865824] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] There are 64 instances to clean {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11679}} [ 1083.865980] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 9b48b942-5e29-4fea-83b7-a1551f78345c] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1083.955213] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.955543] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17daa68d-6f87-4274-b5ca-7d4d66524c6e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.962814] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1083.962814] env[61964]: value = "task-1041608" [ 1083.962814] env[61964]: _type = "Task" [ 1083.962814] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.970642] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.180462] env[61964]: DEBUG oslo_concurrency.lockutils [req-35fb1206-1b79-4293-a4fa-ada3be64974d req-9e63aa1e-f30f-4a78-8f4a-d9c6716f08b7 service nova] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.307252] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527c9d93-7cd0-f29c-daf7-5561a22a0c1d, 'name': SearchDatastore_Task, 'duration_secs': 0.007902} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.308038] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fde1e42-085a-49e2-9413-bc4ef0f41be6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.313046] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1084.313046] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f97a67-d0cc-354f-b2ef-bf871a133b5f" [ 1084.313046] env[61964]: _type = "Task" [ 1084.313046] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.320372] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f97a67-d0cc-354f-b2ef-bf871a133b5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.369067] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6ad71a14-c3d3-4db3-968a-d51d51d86496] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1084.473317] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1084.473549] env[61964]: DEBUG nova.compute.manager [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1084.474321] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cda8c1-2987-4ced-908b-304a5a578f0f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.479849] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.480029] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.480345] env[61964]: DEBUG nova.network.neutron [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1084.823632] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52f97a67-d0cc-354f-b2ef-bf871a133b5f, 'name': SearchDatastore_Task, 'duration_secs': 0.009125} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.823900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.824185] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1084.824443] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3feb448b-a6d8-4703-b9f4-386ba05b078f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.831483] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1084.831483] env[61964]: value = "task-1041610" [ 1084.831483] env[61964]: _type = "Task" [ 1084.831483] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.839172] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.871947] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 204fc3b1-c4a6-43ef-a408-e41be883034e] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1085.243959] env[61964]: DEBUG nova.network.neutron [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.341317] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473522} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.341581] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1085.341806] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.342105] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f0e2b50-06cf-4847-9138-caf3bb1054e9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.348573] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1085.348573] env[61964]: value = "task-1041611" [ 1085.348573] env[61964]: _type = "Task" [ 1085.348573] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.357164] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.374749] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: ee273d32-a49f-43b3-8b08-8017eb1b1636] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1085.746586] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.857678] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062779} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.857957] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.858724] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb391270-d85d-4a25-b253-0c2e9546a4b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.880713] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.881125] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: ada96855-b35f-4ff5-b3ce-e4fa1baaedc4] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1085.882784] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-765ec321-caaf-492c-b955-2dde7e95903a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.903182] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1085.903182] env[61964]: value = "task-1041612" [ 1085.903182] env[61964]: _type = "Task" [ 1085.903182] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.911442] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.986265] env[61964]: DEBUG nova.compute.manager [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-vif-unplugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1085.986522] env[61964]: DEBUG oslo_concurrency.lockutils [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.986762] env[61964]: DEBUG oslo_concurrency.lockutils [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.986940] env[61964]: DEBUG oslo_concurrency.lockutils [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.987242] env[61964]: DEBUG nova.compute.manager [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] No waiting events found dispatching network-vif-unplugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1085.987430] env[61964]: WARNING nova.compute.manager [req-804c8e9b-51c5-45af-aa48-e2c7c8aa1868 req-7f5eb55b-004d-456c-96d6-47aea88ed8e4 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received unexpected event network-vif-unplugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 for instance with vm_state shelved and task_state shelving_offloading. [ 1086.118836] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1086.119879] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab0482a-2318-44af-99b6-631004f1fe0a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.128196] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1086.128447] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f053bc29-69e4-4acd-98d7-02cfad89fafd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.187870] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1086.188186] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1086.188395] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore1] 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1086.188659] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9098ea80-99e4-4903-a184-60343139c621 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.194850] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1086.194850] env[61964]: value = "task-1041614" [ 1086.194850] env[61964]: _type = "Task" [ 1086.194850] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.202565] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041614, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.399022] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 3dc30299-8462-4714-8e1c-8a10412e84d5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1086.412917] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041612, 'name': ReconfigVM_Task, 'duration_secs': 0.269423} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.413160] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Reconfigured VM instance instance-0000006c to attach disk [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.413774] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c27e2978-d591-43da-bfe8-bfceb93ebbfb {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.419404] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1086.419404] env[61964]: value = "task-1041615" [ 1086.419404] env[61964]: _type = "Task" [ 1086.419404] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.426843] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041615, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.704697] env[61964]: DEBUG oslo_vmware.api [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041614, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139856} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.704963] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.705192] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1086.705376] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1086.726379] env[61964]: INFO nova.scheduler.client.report [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted allocations for instance 2da49b28-0cb0-4486-8e69-ce6fba20387c [ 1086.902735] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0c1e5fc1-4e25-489f-b640-253ce03139d5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1086.929340] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041615, 'name': Rename_Task, 'duration_secs': 0.135123} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.929613] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1086.929847] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da9fb6fa-92f6-4ad6-b531-08a63b7ece3d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.936402] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1086.936402] env[61964]: value = "task-1041616" [ 1086.936402] env[61964]: _type = "Task" [ 1086.936402] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.943363] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.230875] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.231276] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.231569] env[61964]: DEBUG nova.objects.instance [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'resources' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.406082] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: f7f18f7b-5e58-426c-aa5d-7063a50c32ef] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1087.446964] env[61964]: DEBUG oslo_vmware.api [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041616, 'name': PowerOnVM_Task, 'duration_secs': 0.460223} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.447817] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.447817] env[61964]: INFO nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Took 6.71 seconds to spawn the instance on the hypervisor. [ 1087.447817] env[61964]: DEBUG nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1087.448483] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b88ad5a-1d9d-496c-b704-818f3689d73e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.734283] env[61964]: DEBUG nova.objects.instance [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'numa_topology' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.911139] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0da26d16-84f2-4c00-97f5-ff132fa122d2] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1087.966123] env[61964]: INFO nova.compute.manager [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Took 12.46 seconds to build instance. [ 1088.022309] env[61964]: DEBUG nova.compute.manager [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1088.022669] env[61964]: DEBUG nova.compute.manager [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing instance network info cache due to event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1088.022874] env[61964]: DEBUG oslo_concurrency.lockutils [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.024100] env[61964]: DEBUG oslo_concurrency.lockutils [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.024100] env[61964]: DEBUG nova.network.neutron [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1088.237659] env[61964]: DEBUG nova.objects.base [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Object Instance<2da49b28-0cb0-4486-8e69-ce6fba20387c> lazy-loaded attributes: resources,numa_topology {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1088.295275] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576d976d-1c4c-40e6-9d25-f5f8b3dd6cd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.305100] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ad3c52-11a9-46af-8ca1-ad5baef85632 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.343723] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db03fdc-5e71-445d-a97c-bb79c694e956 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.351418] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80293b4-4017-4fcc-b27e-949000308892 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.371273] env[61964]: DEBUG nova.compute.provider_tree [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.415380] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 81f821a3-6238-4ab1-9fe3-56d3f83ad52c] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1088.468137] env[61964]: DEBUG oslo_concurrency.lockutils [None req-8fd47b2c-72ce-45ca-b97d-a4e805f8af37 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.973s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.861890] env[61964]: DEBUG nova.network.neutron [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updated VIF entry in instance network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1088.862289] env[61964]: DEBUG nova.network.neutron [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.873921] env[61964]: DEBUG nova.scheduler.client.report [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1088.918962] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2436ae21-6244-4f8b-abb0-1e7530f07dc5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1088.950312] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.104832] env[61964]: DEBUG nova.compute.manager [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1089.106359] env[61964]: DEBUG nova.compute.manager [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing instance network info cache due to event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1089.106359] env[61964]: DEBUG oslo_concurrency.lockutils [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.106359] env[61964]: DEBUG oslo_concurrency.lockutils [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.106359] env[61964]: DEBUG nova.network.neutron [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1089.365766] env[61964]: DEBUG oslo_concurrency.lockutils [req-589eae50-062c-458e-a18a-661ad86ccf50 req-38bc3596-6eca-493c-ada1-bf2ff0991c41 service nova] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.380843] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.150s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.422395] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 90679b1f-2c9c-4049-bfb9-db573220163b] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1089.811318] env[61964]: DEBUG nova.network.neutron [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updated VIF entry in instance network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1089.811703] env[61964]: DEBUG nova.network.neutron [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.889256] env[61964]: DEBUG oslo_concurrency.lockutils [None req-2543a5af-836f-4d0e-900f-b3642761b26d tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.048s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.890600] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.940s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.890722] env[61964]: INFO nova.compute.manager [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Unshelving [ 1089.925849] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 15e92512-3c92-43f0-891c-84bdaad1d186] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1090.313952] env[61964]: DEBUG oslo_concurrency.lockutils [req-25463983-7a29-40aa-aa8b-61df91610c29 req-f21a364a-a018-4b74-bca2-198c7a3e30a8 service nova] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.428475] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: f983316e-e012-41fd-bf20-7d9cc5accd71] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1090.903988] env[61964]: DEBUG nova.compute.utils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.931946] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 39683f54-fa99-448d-91ab-171397be48aa] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1091.407068] env[61964]: INFO nova.virt.block_device [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Booting with volume eb2582c1-607d-4495-8926-da3b30ced75f at /dev/sdb [ 1091.435085] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: c4fabe85-1b95-4981-9e05-dbfedbbe1e1e] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1091.441168] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bcb8119-0856-4d64-b84c-1028e4631db1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.450706] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51228833-26b8-4101-a9ce-1539bdd8b444 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.474461] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa8ba33f-78d9-4e89-9dac-3a9e2fa8f0f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.481809] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa357821-26cb-45c8-8d40-291ae6634094 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.504584] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19eaa779-176d-4da3-a849-323aa6c6fa41 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.510320] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df4c4f4-e865-45f3-8cce-fe798d57e1a3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.522679] env[61964]: DEBUG nova.virt.block_device [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating existing volume attachment record: 0aa0c341-6359-4e1a-a4d7-9b137556c7ae {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1091.938154] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2c77a07d-c949-4aef-8d6f-132969d58cee] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1092.441397] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 58f1ad74-ea72-45c8-b98c-c85674989342] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1092.944537] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 34e00bc8-8124-4ee9-a0a4-c70bf611996c] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1093.447888] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: d8f24177-b695-4d40-958d-9272b4f683cc] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1093.954019] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 003a5b74-2b8d-4e14-a7ee-db8006f81dfa] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1094.454472] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: f5acabe4-674b-4c4c-85b4-b9270926017b] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1094.957897] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 8a14232f-5cc6-4856-9291-eb0a1bd2ab69] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1095.461025] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 47378856-4cbf-4cf6-aecd-ae935885df3d] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1095.964599] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 4d9c24cf-e11b-4d95-ad41-538d91279689] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1096.467629] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: f4c9bdf0-9e53-4a41-9a3f-02bd9198f7a8] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1096.971337] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: ba0f227f-640a-4765-a84e-61bd81128356] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1097.131524] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.132256] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.132256] env[61964]: DEBUG nova.objects.instance [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'pci_requests' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.474594] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 52f358e4-beb2-4b96-8f96-5cd4da103eda] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1097.636151] env[61964]: DEBUG nova.objects.instance [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'numa_topology' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.978463] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 50b04cbf-9064-4010-b350-638e8096bb96] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1098.139029] env[61964]: INFO nova.compute.claims [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.482384] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a1e5c369-4233-4794-adda-1ba1e89e8154] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1098.985777] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 84821045-27b2-4a99-87f9-988b8615b83f] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1099.200016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d3f6ce-d4f3-4d4e-9d4f-15e3d49bc611 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.208502] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495f71d8-a111-4980-b8d9-37e77487b5b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.240057] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0d5f80-9cd5-41c8-8bb3-a94ae37e7877 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.248152] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd05e00-4a0d-4fc9-8c44-bae2f82de3f1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.261791] env[61964]: DEBUG nova.compute.provider_tree [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.489611] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: fb96bf3a-ea2c-414d-bb29-eca2b07d51b6] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1099.764708] env[61964]: DEBUG nova.scheduler.client.report [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1099.992862] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: eb97402e-e722-4cc3-a1d3-1b360ab1e1a7] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1100.270272] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.138s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.310088] env[61964]: INFO nova.network.neutron [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating port d7fbf12c-2a5d-4e7f-a51e-b087b2088571 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1100.496971] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 93720ce5-834a-4d8d-b230-13d38620b688] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1101.000988] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e2c62477-e5fd-4025-9c1c-5e7be9dc4ac0] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1101.504835] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: b90e7ebd-8dc6-495b-99bf-15ff24763af4] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1101.720606] env[61964]: DEBUG nova.compute.manager [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1101.720843] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.721071] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.721251] env[61964]: DEBUG oslo_concurrency.lockutils [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.721431] env[61964]: DEBUG nova.compute.manager [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] No waiting events found dispatching network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1101.721606] env[61964]: WARNING nova.compute.manager [req-dc25244d-af76-4fda-b65c-c78c5d4e187a req-5accc594-5dd2-4e68-ba3c-8be967373c68 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received unexpected event network-vif-plugged-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 for instance with vm_state shelved_offloaded and task_state spawning. [ 1101.809325] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.809526] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.809722] env[61964]: DEBUG nova.network.neutron [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1102.007938] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 138f44e4-e12e-4f89-a9b2-8a512b53cdf5] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1102.506818] env[61964]: DEBUG nova.network.neutron [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.510739] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 8bf586a2-0b56-4f5a-9f33-d537fba00d32] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1103.009299] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.013866] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 2a185e4b-c636-418a-a75e-2cf158f550c0] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1103.036711] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9a192b3cd6ba458d0605da6b5a728ea6',container_format='bare',created_at=2024-09-30T16:29:11Z,direct_url=,disk_format='vmdk',id=92b8e67a-6c74-49a2-b4f1-60044fd1bb46,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1292315013-shelved',owner='4d63f8afe09843efb54de6ea85f276aa',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-09-30T16:29:30Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1103.036973] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.037154] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1103.037345] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.037516] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1103.037689] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1103.037899] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1103.038076] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1103.038251] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1103.038417] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1103.038596] env[61964]: DEBUG nova.virt.hardware [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1103.039478] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5aa7db-eeff-4a2d-8f1a-c1421126bb5e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.048031] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31b0ca4-b101-41fd-9fbd-a3113dee9d03 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.061321] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:90:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd7fbf12c-2a5d-4e7f-a51e-b087b2088571', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.068632] env[61964]: DEBUG oslo.service.loopingcall [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.068861] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1103.069071] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c298e4f8-a2d0-479b-b34f-864434fcbe01 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.087996] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.087996] env[61964]: value = "task-1041621" [ 1103.087996] env[61964]: _type = "Task" [ 1103.087996] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.094969] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041621, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.518319] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 1c922228-4ba9-43ed-8e5c-3ca15e48ceab] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1103.599283] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041621, 'name': CreateVM_Task, 'duration_secs': 0.29524} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.599425] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1103.600049] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.600226] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.600624] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1103.600876] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3816323d-9faa-4cc2-86d3-0cde8dc0375e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.605139] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1103.605139] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5283425f-6313-cc3e-2e09-e8cc318dd0f5" [ 1103.605139] env[61964]: _type = "Task" [ 1103.605139] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.612140] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5283425f-6313-cc3e-2e09-e8cc318dd0f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.745774] env[61964]: DEBUG nova.compute.manager [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1103.745983] env[61964]: DEBUG nova.compute.manager [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing instance network info cache due to event network-changed-d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1103.746225] env[61964]: DEBUG oslo_concurrency.lockutils [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] Acquiring lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.746374] env[61964]: DEBUG oslo_concurrency.lockutils [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] Acquired lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.746570] env[61964]: DEBUG nova.network.neutron [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Refreshing network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1104.021484] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: ab33fb75-4114-4a26-8692-befb407b7460] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1104.115249] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.115507] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Processing image 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.115747] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.115900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.116098] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.116346] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-552fbaf7-2655-463f-8561-19d44f3cbf49 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.132864] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.133074] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1104.133755] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0fbf67a-3c7b-4618-adbd-39661baeeb57 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.138608] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1104.138608] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52969c87-1895-f656-a3d1-c01a0f3b6757" [ 1104.138608] env[61964]: _type = "Task" [ 1104.138608] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.145616] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52969c87-1895-f656-a3d1-c01a0f3b6757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.449224] env[61964]: DEBUG nova.network.neutron [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updated VIF entry in instance network info cache for port d7fbf12c-2a5d-4e7f-a51e-b087b2088571. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.449600] env[61964]: DEBUG nova.network.neutron [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [{"id": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "address": "fa:16:3e:70:90:43", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7fbf12c-2a", "ovs_interfaceid": "d7fbf12c-2a5d-4e7f-a51e-b087b2088571", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.524144] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a23e7f40-1b3e-4c8f-b664-30f6314923a1] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1104.648633] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1104.648853] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Fetch image to [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8/OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1104.649056] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Downloading stream optimized image 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 to [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8/OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8.vmdk on the data store datastore2 as vApp {{(pid=61964) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1104.649243] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Downloading image file data 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 to the ESX as VM named 'OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8' {{(pid=61964) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1104.713341] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1104.713341] env[61964]: value = "resgroup-9" [ 1104.713341] env[61964]: _type = "ResourcePool" [ 1104.713341] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1104.713622] env[61964]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0a672ce0-b7b9-4fe7-b922-53b55a8225c3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.733420] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease: (returnval){ [ 1104.733420] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528b64b9-bef8-6b12-010f-b9b711c0acd4" [ 1104.733420] env[61964]: _type = "HttpNfcLease" [ 1104.733420] env[61964]: } obtained for vApp import into resource pool (val){ [ 1104.733420] env[61964]: value = "resgroup-9" [ 1104.733420] env[61964]: _type = "ResourcePool" [ 1104.733420] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1104.733718] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the lease: (returnval){ [ 1104.733718] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528b64b9-bef8-6b12-010f-b9b711c0acd4" [ 1104.733718] env[61964]: _type = "HttpNfcLease" [ 1104.733718] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1104.739462] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1104.739462] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528b64b9-bef8-6b12-010f-b9b711c0acd4" [ 1104.739462] env[61964]: _type = "HttpNfcLease" [ 1104.739462] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1104.952747] env[61964]: DEBUG oslo_concurrency.lockutils [req-546af89f-283d-4180-b564-68b5218fdcf5 req-0b1d9eae-5a4c-401b-977e-21ffa4d70a56 service nova] Releasing lock "refresh_cache-2da49b28-0cb0-4486-8e69-ce6fba20387c" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.027711] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 560bd013-eae1-4541-aaa2-0732d8f20e4a] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1105.241342] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1105.241342] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528b64b9-bef8-6b12-010f-b9b711c0acd4" [ 1105.241342] env[61964]: _type = "HttpNfcLease" [ 1105.241342] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1105.241585] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1105.241585] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]528b64b9-bef8-6b12-010f-b9b711c0acd4" [ 1105.241585] env[61964]: _type = "HttpNfcLease" [ 1105.241585] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1105.242313] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4569f2ac-3cb7-432c-95b3-f1276eb25f71 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.249215] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1105.249397] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1105.312141] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e7167783-7929-4106-9548-44dff087b6b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.531303] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 062da016-f4d6-424d-90e6-5903940e2daa] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1106.035963] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: beddafaf-8a91-4bfa-af3c-fa69977677ba] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1106.355090] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1106.355372] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1106.356421] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f523bd-db07-47a8-8dab-f9be9d4040f5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.363491] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1106.363665] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1106.363949] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-571505d2-fd9e-4837-814c-3eae53643faa {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.540737] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6a03c9cc-bf79-4d60-98fd-48b2e5bc547d] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1106.567188] env[61964]: DEBUG oslo_vmware.rw_handles [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bb30b0-62a1-9ca6-dbb6-c88670ff5ac8/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1106.567536] env[61964]: INFO nova.virt.vmwareapi.images [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Downloaded image file data 92b8e67a-6c74-49a2-b4f1-60044fd1bb46 [ 1106.569065] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ff51a-2c4e-47b2-b7f6-7da6fb7939a6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.586318] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7912dc64-35ca-4233-a8f5-4b2ceb67d474 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.606983] env[61964]: INFO nova.virt.vmwareapi.images [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] The imported VM was unregistered [ 1106.609399] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1106.609651] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1106.609921] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5f1790c-d99e-4889-a91b-6ae3e2876674 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.619657] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46 {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1106.619844] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8/OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8.vmdk to [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk. {{(pid=61964) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1106.620178] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7b369226-e1a5-459a-b87e-c5833793a5c0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.626491] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1106.626491] env[61964]: value = "task-1041624" [ 1106.626491] env[61964]: _type = "Task" [ 1106.626491] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.633826] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.869624] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.869966] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.044581] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 595c94f1-9dec-454d-b301-5656287f53de] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1107.136612] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.374352] env[61964]: DEBUG nova.compute.utils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.548396] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: e93f98b3-4763-4f02-abd3-c24a9ab8dc17] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1107.638469] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.878468] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.052232] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 9adedaa2-8594-4ddd-8f85-a41174207ef8] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1108.138762] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.555814] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6b2b44b9-ce10-43e4-82a7-a9cfa67a0ba4] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1108.640066] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.949431] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.949647] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.949913] env[61964]: INFO nova.compute.manager [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Attaching volume f45f6103-0a46-41ff-9452-f33aa585d5ed to /dev/sdb [ 1108.979560] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f14c1e3-4d71-4015-ad23-f9a1e2191f0e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.986789] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9480b933-87ac-4e23-ad5b-61b9cf0523f2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.000215] env[61964]: DEBUG nova.virt.block_device [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating existing volume attachment record: 46cb5a32-a2e1-466a-a22b-de37fcb02f2c {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1109.059808] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 0156727e-60dc-4e24-99e6-96b3e735aa0e] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1109.139289] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041624, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.16312} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.139289] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8/OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8.vmdk to [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk. [ 1109.139289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Cleaning up location [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1109.139289] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_c29842de-3185-4036-ac59-cf53f01f7cb8 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.139496] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d17851c-0df6-4b49-aae6-7e56777641ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.145594] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1109.145594] env[61964]: value = "task-1041625" [ 1109.145594] env[61964]: _type = "Task" [ 1109.145594] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.153162] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041625, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.563154] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: fb538d67-47c3-4639-a642-b3358f97df31] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1109.655642] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041625, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036523} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.655918] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.656110] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.656389] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk to [datastore2] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1109.656653] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a25e03c9-e59e-40d4-bec7-314475631485 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.662228] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1109.662228] env[61964]: value = "task-1041627" [ 1109.662228] env[61964]: _type = "Task" [ 1109.662228] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.669480] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.066095] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 58f8d368-fa0d-4c8d-9009-e6e1c5940032] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1110.172341] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.569353] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 4104dee9-34db-4396-9189-8dc677d4b677] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1110.673012] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.073426] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 46435c68-f85a-4360-b2b5-6296afc33c3e] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1111.174273] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.577657] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 6ee845c3-5f79-4704-8b7b-cd3770202647] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1111.674636] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.080988] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: df0a9a69-bd44-4da4-ba3a-9ba241c010a4] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1112.173963] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041627, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.235901} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.174248] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/92b8e67a-6c74-49a2-b4f1-60044fd1bb46/92b8e67a-6c74-49a2-b4f1-60044fd1bb46.vmdk to [datastore2] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1112.174988] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f05bbb-bbd3-4a8b-81fd-594b19e341b4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.196062] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.196331] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b5c6e06-3bc0-43df-a5f0-87c5419833fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.214662] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1112.214662] env[61964]: value = "task-1041629" [ 1112.214662] env[61964]: _type = "Task" [ 1112.214662] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.222176] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.584617] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: af413b1d-9e7e-43d0-a39a-fb9dda9cf281] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1112.725044] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041629, 'name': ReconfigVM_Task, 'duration_secs': 0.268567} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.725044] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 2da49b28-0cb0-4486-8e69-ce6fba20387c/2da49b28-0cb0-4486-8e69-ce6fba20387c.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.726126] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encrypted': False, 'device_name': '/dev/sda', 'guest_format': None, 'disk_bus': None, 'encryption_secret_uuid': None, 'device_type': 'disk', 'encryption_format': None, 'size': 0, 'encryption_options': None, 'boot_index': 0, 'image_id': '5f28385d-6ea0-420d-8a26-4cb693714c14'}], 'ephemerals': [], 'block_device_mapping': [{'mount_device': '/dev/sdb', 'delete_on_termination': False, 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': '0aa0c341-6359-4e1a-a4d7-9b137556c7ae', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230578', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'name': 'volume-eb2582c1-607d-4495-8926-da3b30ced75f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '2da49b28-0cb0-4486-8e69-ce6fba20387c', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'serial': 'eb2582c1-607d-4495-8926-da3b30ced75f'}, 'volume_type': None}], 'swap': None} {{(pid=61964) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1112.726373] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1112.726589] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230578', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'name': 'volume-eb2582c1-607d-4495-8926-da3b30ced75f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '2da49b28-0cb0-4486-8e69-ce6fba20387c', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'serial': 'eb2582c1-607d-4495-8926-da3b30ced75f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1112.727367] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66b0155-3ac4-4657-92ac-723f07df011b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.742474] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3af266-dc7c-44b6-9e49-6d61d3651750 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.765576] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] volume-eb2582c1-607d-4495-8926-da3b30ced75f/volume-eb2582c1-607d-4495-8926-da3b30ced75f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.765847] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03661f66-e613-4392-9f1c-8c8a06f45619 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.783203] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1112.783203] env[61964]: value = "task-1041630" [ 1112.783203] env[61964]: _type = "Task" [ 1112.783203] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.792236] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.087932] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 16276be9-b305-4d1d-afde-bc98be42687a] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1113.292517] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041630, 'name': ReconfigVM_Task, 'duration_secs': 0.295967} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.292902] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfigured VM instance instance-00000064 to attach disk [datastore2] volume-eb2582c1-607d-4495-8926-da3b30ced75f/volume-eb2582c1-607d-4495-8926-da3b30ced75f.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.297556] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b59314a6-3394-451d-a8ef-83a2323af249 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.311200] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1113.311200] env[61964]: value = "task-1041631" [ 1113.311200] env[61964]: _type = "Task" [ 1113.311200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.318525] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041631, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.544669] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1113.544934] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230581', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'name': 'volume-f45f6103-0a46-41ff-9452-f33aa585d5ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '738ae769-4ebd-4828-9414-1ba8309833d6', 'attached_at': '', 'detached_at': '', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'serial': 'f45f6103-0a46-41ff-9452-f33aa585d5ed'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1113.545844] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aef96c0-65d4-4e7c-bcb4-0fea2c36edcc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.561203] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a947ef6-f5ec-4613-83bb-275b0ec4c020 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.584435] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-f45f6103-0a46-41ff-9452-f33aa585d5ed/volume-f45f6103-0a46-41ff-9452-f33aa585d5ed.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.584710] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55e4cea0-1a10-407c-8821-ff8dd2a3e96c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.597508] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: aef4c3a7-641a-4356-9187-ae4c082ccde9] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1113.603842] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1113.603842] env[61964]: value = "task-1041632" [ 1113.603842] env[61964]: _type = "Task" [ 1113.603842] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.611645] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.820953] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041631, 'name': ReconfigVM_Task, 'duration_secs': 0.170094} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.821230] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230578', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'name': 'volume-eb2582c1-607d-4495-8926-da3b30ced75f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '2da49b28-0cb0-4486-8e69-ce6fba20387c', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'serial': 'eb2582c1-607d-4495-8926-da3b30ced75f'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1113.821813] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fe6112e-ae51-455f-822a-8d6e09070d48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.827804] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1113.827804] env[61964]: value = "task-1041633" [ 1113.827804] env[61964]: _type = "Task" [ 1113.827804] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.835265] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041633, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.100768] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: cdfc0bf1-d603-4580-8527-8b06e5ae0799] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1114.112988] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041632, 'name': ReconfigVM_Task, 'duration_secs': 0.344325} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.113303] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-f45f6103-0a46-41ff-9452-f33aa585d5ed/volume-f45f6103-0a46-41ff-9452-f33aa585d5ed.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.118013] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd5ea35e-6c49-4b87-8791-1c8de6fca163 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.133550] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1114.133550] env[61964]: value = "task-1041634" [ 1114.133550] env[61964]: _type = "Task" [ 1114.133550] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.142852] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041634, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.338028] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041633, 'name': Rename_Task, 'duration_secs': 0.161711} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.338347] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1114.338634] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0e3fb0b-7809-4c84-8a29-a6f041ab4bf1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.344387] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1114.344387] env[61964]: value = "task-1041635" [ 1114.344387] env[61964]: _type = "Task" [ 1114.344387] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.352741] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.604397] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 65566c20-6d69-471c-b098-3c30c01d9955] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1114.643158] env[61964]: DEBUG oslo_vmware.api [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041634, 'name': ReconfigVM_Task, 'duration_secs': 0.137667} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.643480] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230581', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'name': 'volume-f45f6103-0a46-41ff-9452-f33aa585d5ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '738ae769-4ebd-4828-9414-1ba8309833d6', 'attached_at': '', 'detached_at': '', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'serial': 'f45f6103-0a46-41ff-9452-f33aa585d5ed'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1114.854316] env[61964]: DEBUG oslo_vmware.api [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041635, 'name': PowerOnVM_Task, 'duration_secs': 0.418316} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.854616] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1114.950465] env[61964]: DEBUG nova.compute.manager [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1114.951444] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73d8da1-6061-42bb-bbed-2e28cc93c2f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.107981] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a995ff8c-75b9-4089-ad4c-9e6baa33fe8c] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1115.467465] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c0005130-79c7-447c-bf3c-914574b62712 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.577s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.610912] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 63793ce6-d511-403e-8a4b-cad8c4157449] Instance has had 0 of 5 cleanup attempts {{(pid=61964) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11683}} [ 1115.682746] env[61964]: DEBUG nova.objects.instance [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 738ae769-4ebd-4828-9414-1ba8309833d6 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.187704] env[61964]: DEBUG oslo_concurrency.lockutils [None req-85680e15-3af0-4db7-a3ab-81f4f204cacc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.238s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.474543] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.474819] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.978353] env[61964]: INFO nova.compute.manager [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Detaching volume f45f6103-0a46-41ff-9452-f33aa585d5ed [ 1117.011550] env[61964]: INFO nova.virt.block_device [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Attempting to driver detach volume f45f6103-0a46-41ff-9452-f33aa585d5ed from mountpoint /dev/sdb [ 1117.011803] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1117.012010] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230581', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'name': 'volume-f45f6103-0a46-41ff-9452-f33aa585d5ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '738ae769-4ebd-4828-9414-1ba8309833d6', 'attached_at': '', 'detached_at': '', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'serial': 'f45f6103-0a46-41ff-9452-f33aa585d5ed'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1117.012915] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fbf041-3e2f-4815-a07a-27709971eedf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.035478] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c851cea-ca9e-454e-b945-9db201fd9e7f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.041955] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06098811-e86b-4b09-8912-b02dd3f0e56f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.061108] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a48895-82be-4b8d-8f88-b2e7e51febd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.075018] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] The volume has not been displaced from its original location: [datastore2] volume-f45f6103-0a46-41ff-9452-f33aa585d5ed/volume-f45f6103-0a46-41ff-9452-f33aa585d5ed.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1117.080274] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.080569] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dacd2635-fce6-40ea-88ab-c6530f8718b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.098120] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1117.098120] env[61964]: value = "task-1041636" [ 1117.098120] env[61964]: _type = "Task" [ 1117.098120] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.107566] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041636, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.607949] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041636, 'name': ReconfigVM_Task, 'duration_secs': 0.211602} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.608274] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1117.612785] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-524c3471-fdcf-45a1-8bae-09e96450f268 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.627616] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1117.627616] env[61964]: value = "task-1041637" [ 1117.627616] env[61964]: _type = "Task" [ 1117.627616] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.635373] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041637, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.137201] env[61964]: DEBUG oslo_vmware.api [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041637, 'name': ReconfigVM_Task, 'duration_secs': 0.138682} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.137553] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230581', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'name': 'volume-f45f6103-0a46-41ff-9452-f33aa585d5ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '738ae769-4ebd-4828-9414-1ba8309833d6', 'attached_at': '', 'detached_at': '', 'volume_id': 'f45f6103-0a46-41ff-9452-f33aa585d5ed', 'serial': 'f45f6103-0a46-41ff-9452-f33aa585d5ed'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1118.676655] env[61964]: DEBUG nova.objects.instance [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 738ae769-4ebd-4828-9414-1ba8309833d6 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.683974] env[61964]: DEBUG oslo_concurrency.lockutils [None req-6b3f8f73-ef45-4e58-8563-fb59bc4c7fa4 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.209s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.717408] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.717835] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.717895] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.718101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.718283] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.720596] env[61964]: INFO nova.compute.manager [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Terminating instance [ 1121.224915] env[61964]: DEBUG nova.compute.manager [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1121.225207] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1121.226206] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5548d92e-d8cd-4f00-80f5-f9a4e8007561 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.233976] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1121.234222] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-227430cc-83bb-45da-aca1-70e9ddbdac48 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.239778] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1121.239778] env[61964]: value = "task-1041638" [ 1121.239778] env[61964]: _type = "Task" [ 1121.239778] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.247496] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.749654] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041638, 'name': PowerOffVM_Task, 'duration_secs': 0.231493} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.750029] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1121.750029] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1121.750278] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e02665e3-10ef-4ab6-93d2-c7b02b4aea9c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.810873] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1121.811109] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1121.811303] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleting the datastore file [datastore1] 738ae769-4ebd-4828-9414-1ba8309833d6 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.811578] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f41ff95f-81f4-4bbf-98a6-0b1fc8c23055 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.817493] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1121.817493] env[61964]: value = "task-1041640" [ 1121.817493] env[61964]: _type = "Task" [ 1121.817493] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.824713] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.328888] env[61964]: DEBUG oslo_vmware.api [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144633} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.329175] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1122.329368] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1122.329551] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1122.329731] env[61964]: INFO nova.compute.manager [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1122.329997] env[61964]: DEBUG oslo.service.loopingcall [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.330210] env[61964]: DEBUG nova.compute.manager [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1122.330312] env[61964]: DEBUG nova.network.neutron [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1122.832670] env[61964]: DEBUG nova.compute.manager [req-d3a17956-f87c-4c2d-a28a-0406fea1a878 req-340f001c-fcb9-422b-8201-1ad47cd135ec service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Received event network-vif-deleted-e5ee90c2-06be-418d-b67e-9257fb728e47 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1122.832920] env[61964]: INFO nova.compute.manager [req-d3a17956-f87c-4c2d-a28a-0406fea1a878 req-340f001c-fcb9-422b-8201-1ad47cd135ec service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Neutron deleted interface e5ee90c2-06be-418d-b67e-9257fb728e47; detaching it from the instance and deleting it from the info cache [ 1122.833093] env[61964]: DEBUG nova.network.neutron [req-d3a17956-f87c-4c2d-a28a-0406fea1a878 req-340f001c-fcb9-422b-8201-1ad47cd135ec service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.108051] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.307222] env[61964]: DEBUG nova.network.neutron [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.336029] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6c093da-578f-469e-bc05-c9cbe8b814f0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.346016] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c41ec3a-6e93-4eca-a7a5-a30ac3d8d212 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.373129] env[61964]: DEBUG nova.compute.manager [req-d3a17956-f87c-4c2d-a28a-0406fea1a878 req-340f001c-fcb9-422b-8201-1ad47cd135ec service nova] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Detach interface failed, port_id=e5ee90c2-06be-418d-b67e-9257fb728e47, reason: Instance 738ae769-4ebd-4828-9414-1ba8309833d6 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1123.613112] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_power_states {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.809830] env[61964]: INFO nova.compute.manager [-] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] Took 1.48 seconds to deallocate network for instance. [ 1124.118240] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Getting list of instances from cluster (obj){ [ 1124.118240] env[61964]: value = "domain-c8" [ 1124.118240] env[61964]: _type = "ClusterComputeResource" [ 1124.118240] env[61964]: } {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1124.119719] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdeb72c2-cb80-4ed5-8f08-6ea7027c5434 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.131804] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Got total of 2 instances {{(pid=61964) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1124.131958] env[61964]: WARNING nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1124.132119] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Triggering sync for uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10780}} [ 1124.132338] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Triggering sync for uuid 738ae769-4ebd-4828-9414-1ba8309833d6 {{(pid=61964) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10780}} [ 1124.132643] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Triggering sync for uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10780}} [ 1124.132978] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.133222] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.133487] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "738ae769-4ebd-4828-9414-1ba8309833d6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.133706] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.133891] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.134705] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbb6d96-17e5-4dd1-954c-f80d9be77804 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.137598] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca9e1a56-9a1e-48a9-b4d0-8eb7f45dfb5c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.316865] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.317221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.317490] env[61964]: DEBUG nova.objects.instance [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'resources' on Instance uuid 738ae769-4ebd-4828-9414-1ba8309833d6 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.645961] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.648545] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.515s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.868746] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e66abd-1fb0-4bb3-b1e3-3d90e6ea281f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.876110] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0478ef1-38ac-496c-8294-8bd2302a786d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.904661] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35149242-65d7-4112-903b-4565afb38ce4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.910993] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c98545-e90e-4f4d-880c-dc83d0d39cd9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.924422] env[61964]: DEBUG nova.compute.provider_tree [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.427236] env[61964]: DEBUG nova.scheduler.client.report [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1125.932472] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.939057] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.939319] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.939504] env[61964]: INFO nova.compute.manager [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Shelving [ 1125.953074] env[61964]: INFO nova.scheduler.client.report [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted allocations for instance 738ae769-4ebd-4828-9414-1ba8309833d6 [ 1126.460385] env[61964]: DEBUG oslo_concurrency.lockutils [None req-10e98b09-5a1e-42f2-940c-9fb081809ba8 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.743s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.461229] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.328s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.461428] env[61964]: INFO nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: 738ae769-4ebd-4828-9414-1ba8309833d6] During sync_power_state the instance has a pending task (deleting). Skip. [ 1126.461608] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "738ae769-4ebd-4828-9414-1ba8309833d6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.949573] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1126.950716] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-929c14e2-f072-4d9b-8e13-d039d94ecd43 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.957896] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1126.957896] env[61964]: value = "task-1041641" [ 1126.957896] env[61964]: _type = "Task" [ 1126.957896] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.966336] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041641, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.467538] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041641, 'name': PowerOffVM_Task, 'duration_secs': 0.168357} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.467844] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1127.468651] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa69951-42b3-41f5-b992-e80e4b58de87 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.486357] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c774883-8557-4f9f-a358-4449dfa6e811 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.996688] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1127.997199] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0ab0a7d4-b418-463d-822b-6e34b5bd4112 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.005200] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1128.005200] env[61964]: value = "task-1041642" [ 1128.005200] env[61964]: _type = "Task" [ 1128.005200] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.013826] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041642, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.515444] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041642, 'name': CreateSnapshot_Task, 'duration_secs': 0.404188} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.515715] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1128.516492] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d900dd6-1c1a-4d61-a212-bef0a2597340 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.755052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.755052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.033536] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1129.033853] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4310e713-86de-440b-8dc7-e38aaac2d7c1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.042081] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1129.042081] env[61964]: value = "task-1041643" [ 1129.042081] env[61964]: _type = "Task" [ 1129.042081] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.050177] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041643, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.257148] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1129.552767] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041643, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.779571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.779571] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.781145] env[61964]: INFO nova.compute.claims [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1130.052927] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041643, 'name': CloneVM_Task} progress is 100%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.553741] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041643, 'name': CloneVM_Task, 'duration_secs': 1.013396} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.554115] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Created linked-clone VM from snapshot [ 1130.554683] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faebe896-acdf-44cf-8e8c-455d35a16fac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.562936] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Uploading image 0d2132dd-40ec-4b44-b509-6f1ee15d96bf {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1130.586645] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1130.586645] env[61964]: value = "vm-230583" [ 1130.586645] env[61964]: _type = "VirtualMachine" [ 1130.586645] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1130.586888] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6d29cce5-db24-429c-92a8-39af6d369696 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.593253] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease: (returnval){ [ 1130.593253] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c79e8-0a14-ec0d-4e36-8cba8a3ac23f" [ 1130.593253] env[61964]: _type = "HttpNfcLease" [ 1130.593253] env[61964]: } obtained for exporting VM: (result){ [ 1130.593253] env[61964]: value = "vm-230583" [ 1130.593253] env[61964]: _type = "VirtualMachine" [ 1130.593253] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1130.593524] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the lease: (returnval){ [ 1130.593524] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c79e8-0a14-ec0d-4e36-8cba8a3ac23f" [ 1130.593524] env[61964]: _type = "HttpNfcLease" [ 1130.593524] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1130.599245] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1130.599245] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c79e8-0a14-ec0d-4e36-8cba8a3ac23f" [ 1130.599245] env[61964]: _type = "HttpNfcLease" [ 1130.599245] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1130.834779] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21950306-411f-4315-b8a1-7978857b7d26 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.842103] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66ee3d0-3968-469f-a35c-0810e5fbc330 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.870920] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9d54bd-91b9-4643-a9c8-f1563c756884 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.877537] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bbb0f8-ac21-491b-9011-24a6f88b589a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.890834] env[61964]: DEBUG nova.compute.provider_tree [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.101777] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1131.101777] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c79e8-0a14-ec0d-4e36-8cba8a3ac23f" [ 1131.101777] env[61964]: _type = "HttpNfcLease" [ 1131.101777] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1131.102104] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1131.102104] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]526c79e8-0a14-ec0d-4e36-8cba8a3ac23f" [ 1131.102104] env[61964]: _type = "HttpNfcLease" [ 1131.102104] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1131.102813] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d305709-5da2-47e5-8a6c-167d09632cd6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.109902] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1131.110093] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1131.196457] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1f1c9fbc-e7e1-4961-803f-ffbd32d21989 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.394226] env[61964]: DEBUG nova.scheduler.client.report [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1131.899546] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.900193] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1132.405395] env[61964]: DEBUG nova.compute.utils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1132.407029] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1132.407203] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1132.457128] env[61964]: DEBUG nova.policy [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf85b27f46c94956982ca3c713aef9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b79a033963e04e02a98df18f8429263f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1132.761492] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Successfully created port: 6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1132.910245] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1133.920331] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1133.946256] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1133.946560] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.946732] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1133.946927] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.947149] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1133.947320] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1133.947530] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1133.947694] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1133.947865] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1133.948045] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1133.948227] env[61964]: DEBUG nova.virt.hardware [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1133.949575] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ef438c-d0ce-43bc-8885-487d1420ade8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.957338] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c982a1a6-bdc5-46b5-b2d8-41626eebd371 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.245165] env[61964]: DEBUG nova.compute.manager [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Received event network-vif-plugged-6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1134.245433] env[61964]: DEBUG oslo_concurrency.lockutils [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.245724] env[61964]: DEBUG oslo_concurrency.lockutils [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.245940] env[61964]: DEBUG oslo_concurrency.lockutils [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.246151] env[61964]: DEBUG nova.compute.manager [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] No waiting events found dispatching network-vif-plugged-6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1134.246330] env[61964]: WARNING nova.compute.manager [req-71e6a6c0-a745-4f95-aa6d-e40b84b41890 req-4260db91-ecdf-43a4-88f6-e695443a0d77 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Received unexpected event network-vif-plugged-6f4ceb27-b4c3-4358-8596-ac8c27674bdf for instance with vm_state building and task_state spawning. [ 1134.335451] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Successfully updated port: 6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1134.839063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.839063] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.839063] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1135.287186] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.371506] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1135.511154] env[61964]: DEBUG nova.network.neutron [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating instance_info_cache with network_info: [{"id": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "address": "fa:16:3e:b4:87:e2", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f4ceb27-b4", "ovs_interfaceid": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.765633] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.765943] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.766119] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1136.014214] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.014612] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Instance network_info: |[{"id": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "address": "fa:16:3e:b4:87:e2", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f4ceb27-b4", "ovs_interfaceid": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1136.015176] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:87:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97b68ed7-8461-4345-b064-96a1dde53a86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f4ceb27-b4c3-4358-8596-ac8c27674bdf', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1136.022906] env[61964]: DEBUG oslo.service.loopingcall [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.023225] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1136.023529] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b43a3571-674a-40ab-8b8b-04ab538fa479 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.044389] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1136.044389] env[61964]: value = "task-1041645" [ 1136.044389] env[61964]: _type = "Task" [ 1136.044389] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.052929] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041645, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.312151] env[61964]: DEBUG nova.compute.manager [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Received event network-changed-6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1136.312511] env[61964]: DEBUG nova.compute.manager [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Refreshing instance network info cache due to event network-changed-6f4ceb27-b4c3-4358-8596-ac8c27674bdf. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1136.312511] env[61964]: DEBUG oslo_concurrency.lockutils [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] Acquiring lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.312625] env[61964]: DEBUG oslo_concurrency.lockutils [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] Acquired lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.312770] env[61964]: DEBUG nova.network.neutron [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Refreshing network info cache for port 6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1136.557047] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041645, 'name': CreateVM_Task, 'duration_secs': 0.349521} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.557253] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1136.558266] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.558481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.558935] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1136.559284] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2073be2-e676-4e57-ae99-ea84c2dba4c6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.564743] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1136.564743] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52687cae-dca9-6ab3-000e-3a1421cb0a04" [ 1136.564743] env[61964]: _type = "Task" [ 1136.564743] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.574907] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52687cae-dca9-6ab3-000e-3a1421cb0a04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.761503] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.765208] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.015483] env[61964]: DEBUG nova.network.neutron [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updated VIF entry in instance network info cache for port 6f4ceb27-b4c3-4358-8596-ac8c27674bdf. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1137.015904] env[61964]: DEBUG nova.network.neutron [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating instance_info_cache with network_info: [{"id": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "address": "fa:16:3e:b4:87:e2", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f4ceb27-b4", "ovs_interfaceid": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.075242] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52687cae-dca9-6ab3-000e-3a1421cb0a04, 'name': SearchDatastore_Task, 'duration_secs': 0.013177} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.075567] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.075853] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1137.076101] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.076266] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.076455] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1137.076743] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e22135fa-ba26-45de-a025-449068a61eed {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.090435] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1137.090633] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1137.091425] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9363078-7a28-4048-823c-55a98980c116 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.097125] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1137.097125] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5276e237-0a1c-2dab-d2f3-263c669fa9ff" [ 1137.097125] env[61964]: _type = "Task" [ 1137.097125] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.105547] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5276e237-0a1c-2dab-d2f3-263c669fa9ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.518570] env[61964]: DEBUG oslo_concurrency.lockutils [req-8a45b37a-4c8a-415c-902d-7e31a65e2074 req-e9781e37-c7a3-4d67-a7bc-625c02455a03 service nova] Releasing lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.607532] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5276e237-0a1c-2dab-d2f3-263c669fa9ff, 'name': SearchDatastore_Task, 'duration_secs': 0.012351} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.608352] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5284ecfd-a4eb-4527-9dc0-307e6b43d4f6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.613529] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1137.613529] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2f814-55ca-38b2-c88d-3cd8837bec9f" [ 1137.613529] env[61964]: _type = "Task" [ 1137.613529] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.621411] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2f814-55ca-38b2-c88d-3cd8837bec9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.766325] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.766568] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.125624] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a2f814-55ca-38b2-c88d-3cd8837bec9f, 'name': SearchDatastore_Task, 'duration_secs': 0.022475} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.125908] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.126192] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 85355640-f271-4e0a-bf75-0bbcb8487d6f/85355640-f271-4e0a-bf75-0bbcb8487d6f.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1138.126518] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6303923-b289-41e6-a316-107aaa7af319 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.132888] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1138.132888] env[61964]: value = "task-1041646" [ 1138.132888] env[61964]: _type = "Task" [ 1138.132888] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.140383] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041646, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.642722] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041646, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477376} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.643123] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] 85355640-f271-4e0a-bf75-0bbcb8487d6f/85355640-f271-4e0a-bf75-0bbcb8487d6f.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1138.643260] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.643519] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d735002-7784-4b9d-ae01-08dd8697d257 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.650942] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1138.650942] env[61964]: value = "task-1041647" [ 1138.650942] env[61964]: _type = "Task" [ 1138.650942] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.659152] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.766864] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.766864] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1139.161303] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109716} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.161594] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1139.162425] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7523d48-58b3-4eb7-a182-01b5c0c7e25e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.186069] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 85355640-f271-4e0a-bf75-0bbcb8487d6f/85355640-f271-4e0a-bf75-0bbcb8487d6f.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.186334] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bac02fa7-f605-441b-b2c3-c79755dea0f4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.205250] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1139.205250] env[61964]: value = "task-1041648" [ 1139.205250] env[61964]: _type = "Task" [ 1139.205250] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.213358] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.270237] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Didn't find any instances for network info cache update. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10456}} [ 1139.349751] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1139.350769] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d25b92-1a26-4857-8275-1272107a0684 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.357194] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1139.357367] env[61964]: ERROR oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk due to incomplete transfer. [ 1139.357596] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b0d3a33a-564a-4097-ae7b-f0e569803374 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.363703] env[61964]: DEBUG oslo_vmware.rw_handles [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5228d61c-e9cd-9654-de76-47a8fcf8e761/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1139.363886] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Uploaded image 0d2132dd-40ec-4b44-b509-6f1ee15d96bf to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1139.366329] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1139.366564] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4ce8581d-3763-4abf-848a-1521ad6ec8bd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.371922] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1139.371922] env[61964]: value = "task-1041649" [ 1139.371922] env[61964]: _type = "Task" [ 1139.371922] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.379805] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041649, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.716064] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041648, 'name': ReconfigVM_Task, 'duration_secs': 0.325853} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.716430] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 85355640-f271-4e0a-bf75-0bbcb8487d6f/85355640-f271-4e0a-bf75-0bbcb8487d6f.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.717030] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51e77c8f-5ee1-4bb7-84d6-89664f9467dd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.724236] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1139.724236] env[61964]: value = "task-1041650" [ 1139.724236] env[61964]: _type = "Task" [ 1139.724236] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.732138] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041650, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.881970] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041649, 'name': Destroy_Task, 'duration_secs': 0.372756} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.882294] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Destroyed the VM [ 1139.882572] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1139.882883] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4efcba77-96b6-4d6b-8910-302d1f9623e0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.888634] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1139.888634] env[61964]: value = "task-1041651" [ 1139.888634] env[61964]: _type = "Task" [ 1139.888634] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.896795] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041651, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.234550] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041650, 'name': Rename_Task, 'duration_secs': 0.187987} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.234715] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1140.235420] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cfc21cb-53db-4981-8786-9939f52a222b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.240447] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1140.240447] env[61964]: value = "task-1041652" [ 1140.240447] env[61964]: _type = "Task" [ 1140.240447] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.247658] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.398386] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041651, 'name': RemoveSnapshot_Task, 'duration_secs': 0.360402} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.398667] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1140.398936] env[61964]: DEBUG nova.compute.manager [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1140.399708] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74c721b-8844-4e94-9b48-757325ebc2c9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.750171] env[61964]: DEBUG oslo_vmware.api [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041652, 'name': PowerOnVM_Task, 'duration_secs': 0.480473} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.750636] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1140.750693] env[61964]: INFO nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Took 6.83 seconds to spawn the instance on the hypervisor. [ 1140.750834] env[61964]: DEBUG nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1140.751622] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf4940f-0861-4186-9e8d-ebf2a90e4e93 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.765940] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.911891] env[61964]: INFO nova.compute.manager [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Shelve offloading [ 1141.268249] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.268504] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.268646] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.268800] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1141.269496] env[61964]: INFO nova.compute.manager [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Took 11.51 seconds to build instance. [ 1141.271386] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9cae23-1655-4a74-a534-e2304ac1f1ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.284842] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3571426-ac7f-484e-88c1-44232696ad86 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.299711] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee9cd2e-15e5-4bbb-a8af-a94c9c5b4719 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.306299] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a02ce36-ade1-4384-a7cd-12303ce04531 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.335933] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181214MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1141.335933] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.335933] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.414976] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1141.415376] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35e6a083-233d-4690-afb3-edf1cb987a18 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.422655] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1141.422655] env[61964]: value = "task-1041653" [ 1141.422655] env[61964]: _type = "Task" [ 1141.422655] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.430303] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.777566] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c1365539-4835-4722-a174-b247385a87dc tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.022s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.862168] env[61964]: DEBUG nova.compute.manager [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Received event network-changed-6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1141.862351] env[61964]: DEBUG nova.compute.manager [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Refreshing instance network info cache due to event network-changed-6f4ceb27-b4c3-4358-8596-ac8c27674bdf. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1141.862534] env[61964]: DEBUG oslo_concurrency.lockutils [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] Acquiring lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.862682] env[61964]: DEBUG oslo_concurrency.lockutils [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] Acquired lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.862856] env[61964]: DEBUG nova.network.neutron [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Refreshing network info cache for port 6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.932332] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1141.932511] env[61964]: DEBUG nova.compute.manager [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1141.933252] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc7dea9-3924-4d7f-a557-95ca8339784b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.938595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.938765] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.938937] env[61964]: DEBUG nova.network.neutron [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.361468] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1142.361630] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 2da49b28-0cb0-4486-8e69-ce6fba20387c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1142.361755] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance 85355640-f271-4e0a-bf75-0bbcb8487d6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1142.361943] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1142.362105] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1142.409811] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b6eaa0-38e9-4c2f-9c86-51b056bea347 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.417189] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ef5d2c-63d7-49cf-988c-a2111a5f9c94 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.450807] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2387706-53af-4eec-b947-548872598c75 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.458188] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84a6f0e-6aa5-49a1-9681-8e27a96fc267 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.471338] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.591785] env[61964]: DEBUG nova.network.neutron [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updated VIF entry in instance network info cache for port 6f4ceb27-b4c3-4358-8596-ac8c27674bdf. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1142.592169] env[61964]: DEBUG nova.network.neutron [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating instance_info_cache with network_info: [{"id": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "address": "fa:16:3e:b4:87:e2", "network": {"id": "7e615cd9-b033-4a09-b9d4-7976dc15cc95", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1071421161-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b79a033963e04e02a98df18f8429263f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f4ceb27-b4", "ovs_interfaceid": "6f4ceb27-b4c3-4358-8596-ac8c27674bdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.650238] env[61964]: DEBUG nova.network.neutron [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.974471] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1143.095053] env[61964]: DEBUG oslo_concurrency.lockutils [req-a93902c8-e746-45ed-8608-21248acebf08 req-0ff93d57-241f-42a8-a104-751d4e304e8e service nova] Releasing lock "refresh_cache-85355640-f271-4e0a-bf75-0bbcb8487d6f" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.152434] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.472107] env[61964]: DEBUG nova.compute.manager [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-vif-unplugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1143.472378] env[61964]: DEBUG oslo_concurrency.lockutils [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.472604] env[61964]: DEBUG oslo_concurrency.lockutils [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.472782] env[61964]: DEBUG oslo_concurrency.lockutils [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.472959] env[61964]: DEBUG nova.compute.manager [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] No waiting events found dispatching network-vif-unplugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1143.473149] env[61964]: WARNING nova.compute.manager [req-12018e3e-9939-4135-9586-70464ef3df4c req-9e9186ac-7da8-418d-929c-651fda09fc7d service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received unexpected event network-vif-unplugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 for instance with vm_state shelved and task_state shelving_offloading. [ 1143.478515] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1143.478683] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.143s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.548507] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1143.549426] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a10eb04-a57a-4f32-9239-56e03c5b485f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.558932] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1143.559182] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4269abb9-e60c-498f-9658-bbbcd0a0474f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.633239] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1143.633415] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1143.633656] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleting the datastore file [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.633959] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-724e8a7a-101c-4fbc-b574-c1adc544faa5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.641349] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1143.641349] env[61964]: value = "task-1041655" [ 1143.641349] env[61964]: _type = "Task" [ 1143.641349] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.649378] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.151042] env[61964]: DEBUG oslo_vmware.api [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123461} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.151445] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.151490] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1144.151656] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1144.168348] env[61964]: INFO nova.scheduler.client.report [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted allocations for instance cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 [ 1144.673297] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.673604] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.673872] env[61964]: DEBUG nova.objects.instance [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'resources' on Instance uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.176827] env[61964]: DEBUG nova.objects.instance [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'numa_topology' on Instance uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.500072] env[61964]: DEBUG nova.compute.manager [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1145.500351] env[61964]: DEBUG nova.compute.manager [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing instance network info cache due to event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1145.500640] env[61964]: DEBUG oslo_concurrency.lockutils [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.500849] env[61964]: DEBUG oslo_concurrency.lockutils [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.501216] env[61964]: DEBUG nova.network.neutron [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1145.680324] env[61964]: DEBUG nova.objects.base [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1145.723305] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d70d571-47f7-46af-8623-f5edcddfe627 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.730740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e2027d-6fda-4b08-a9d6-440ba2b7a926 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.760862] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea47b97-a8e2-443f-b258-744e85b9ce22 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.767690] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72434dff-095f-46a1-bd48-3cd9a0ac4180 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.780482] env[61964]: DEBUG nova.compute.provider_tree [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.882948] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.208379] env[61964]: DEBUG nova.network.neutron [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updated VIF entry in instance network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1146.208783] env[61964]: DEBUG nova.network.neutron [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": null, "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.283722] env[61964]: DEBUG nova.scheduler.client.report [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.712116] env[61964]: DEBUG oslo_concurrency.lockutils [req-83ebdce8-0141-47d2-9a26-8045a27e340a req-e3b03fd5-91eb-4de8-96bc-959ec924d186 service nova] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.788376] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.115s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.296977] env[61964]: DEBUG oslo_concurrency.lockutils [None req-bddb16bc-d7b5-4b8e-aa8f-7318f56a44b9 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.358s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.297854] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.415s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.298060] env[61964]: INFO nova.compute.manager [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Unshelving [ 1148.326230] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.326570] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.326827] env[61964]: DEBUG nova.objects.instance [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'pci_requests' on Instance uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.830679] env[61964]: DEBUG nova.objects.instance [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'numa_topology' on Instance uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.333425] env[61964]: INFO nova.compute.claims [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1150.392181] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483cbe64-d499-4212-98c0-b7163ae12e64 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.399971] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ab1278-af21-437e-b958-f102069339ec {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.428997] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb2a389-df08-43a1-8b7e-8e8d317ba3f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.435751] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ff1b5b-35bc-4667-9767-5665473baa38 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.449552] env[61964]: DEBUG nova.compute.provider_tree [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.840772] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.841158] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.951870] env[61964]: DEBUG nova.scheduler.client.report [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1151.345331] env[61964]: INFO nova.compute.manager [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Detaching volume eb2582c1-607d-4495-8926-da3b30ced75f [ 1151.376250] env[61964]: INFO nova.virt.block_device [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Attempting to driver detach volume eb2582c1-607d-4495-8926-da3b30ced75f from mountpoint /dev/sdb [ 1151.376510] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1151.376758] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230578', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'name': 'volume-eb2582c1-607d-4495-8926-da3b30ced75f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '2da49b28-0cb0-4486-8e69-ce6fba20387c', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'serial': 'eb2582c1-607d-4495-8926-da3b30ced75f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1151.377663] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6a6885-afda-4a33-a584-640c199c4671 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.399042] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81847fdb-0b46-4089-ba99-ac05b83bb668 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.405922] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc837f9-03fc-40ac-8412-c2a35dc5a8fe {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.425422] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f21eb4d-db85-41ed-a9c9-dc78d3d1f347 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.439906] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] The volume has not been displaced from its original location: [datastore2] volume-eb2582c1-607d-4495-8926-da3b30ced75f/volume-eb2582c1-607d-4495-8926-da3b30ced75f.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1151.444964] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1151.445230] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4512232a-dfb6-4608-9104-8d847092df7a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.457818] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.131s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.464897] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1151.464897] env[61964]: value = "task-1041656" [ 1151.464897] env[61964]: _type = "Task" [ 1151.464897] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.473057] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041656, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.520154] env[61964]: INFO nova.network.neutron [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating port 8d7c51d7-90bb-4a31-8ce4-230157e98433 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1151.974449] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041656, 'name': ReconfigVM_Task, 'duration_secs': 0.233171} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.974654] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1151.979233] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8f9c9cc-50b7-4608-ab9b-0fcd2a809459 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.993522] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1151.993522] env[61964]: value = "task-1041657" [ 1151.993522] env[61964]: _type = "Task" [ 1151.993522] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.000748] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041657, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.503298] env[61964]: DEBUG oslo_vmware.api [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041657, 'name': ReconfigVM_Task, 'duration_secs': 0.133117} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.503657] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230578', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'name': 'volume-eb2582c1-607d-4495-8926-da3b30ced75f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '2da49b28-0cb0-4486-8e69-ce6fba20387c', 'attached_at': '', 'detached_at': '', 'volume_id': 'eb2582c1-607d-4495-8926-da3b30ced75f', 'serial': 'eb2582c1-607d-4495-8926-da3b30ced75f'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1153.044282] env[61964]: DEBUG nova.objects.instance [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'flavor' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.051832] env[61964]: DEBUG oslo_concurrency.lockutils [None req-5b4b7af8-fd4c-4869-af40-b1a0e229723e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.083781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.084190] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.084298] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.084487] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.084664] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.086792] env[61964]: INFO nova.compute.manager [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Terminating instance [ 1155.591536] env[61964]: DEBUG nova.compute.manager [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1155.591813] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1155.592663] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ae08b8-53c1-4e54-b17d-8e441a0e71dc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.601021] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1155.601502] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35696e64-8d99-42e9-b10e-0cd3eccdf7a8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.607284] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1155.607284] env[61964]: value = "task-1041658" [ 1155.607284] env[61964]: _type = "Task" [ 1155.607284] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.616111] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.965824] env[61964]: DEBUG nova.compute.manager [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1155.966064] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.966284] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.966459] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.966635] env[61964]: DEBUG nova.compute.manager [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] No waiting events found dispatching network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1155.966835] env[61964]: WARNING nova.compute.manager [req-1a5e30ce-4ba4-4968-b4a2-1d00ecf0c0d1 req-b8e903c8-2e55-41b2-9214-87419a01f0be service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received unexpected event network-vif-plugged-8d7c51d7-90bb-4a31-8ce4-230157e98433 for instance with vm_state shelved_offloaded and task_state spawning. [ 1156.097238] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.097530] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.097632] env[61964]: DEBUG nova.network.neutron [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1156.116694] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041658, 'name': PowerOffVM_Task, 'duration_secs': 0.180616} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.116957] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1156.117147] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1156.117400] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41c250af-ac2d-42bb-9724-6ed36eed2b33 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.179810] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1156.180057] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1156.180257] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore2] 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1156.180521] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54bca2bd-dfb4-41db-91df-01051b048f40 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.186108] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1156.186108] env[61964]: value = "task-1041660" [ 1156.186108] env[61964]: _type = "Task" [ 1156.186108] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.193440] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.695727] env[61964]: DEBUG oslo_vmware.api [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041660, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135493} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.698071] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.698283] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1156.698472] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1156.698658] env[61964]: INFO nova.compute.manager [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1156.698906] env[61964]: DEBUG oslo.service.loopingcall [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.699127] env[61964]: DEBUG nova.compute.manager [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1156.699223] env[61964]: DEBUG nova.network.neutron [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1156.809402] env[61964]: DEBUG nova.network.neutron [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.156785] env[61964]: DEBUG nova.compute.manager [req-be1d3af8-d414-4d35-ae93-0b6143800f8c req-72a6d3d0-faa0-4e95-8cc4-bc0c9dbe7be7 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Received event network-vif-deleted-d7fbf12c-2a5d-4e7f-a51e-b087b2088571 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1157.157040] env[61964]: INFO nova.compute.manager [req-be1d3af8-d414-4d35-ae93-0b6143800f8c req-72a6d3d0-faa0-4e95-8cc4-bc0c9dbe7be7 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Neutron deleted interface d7fbf12c-2a5d-4e7f-a51e-b087b2088571; detaching it from the instance and deleting it from the info cache [ 1157.157131] env[61964]: DEBUG nova.network.neutron [req-be1d3af8-d414-4d35-ae93-0b6143800f8c req-72a6d3d0-faa0-4e95-8cc4-bc0c9dbe7be7 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.312236] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.341132] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='023758b791847658f44e3fef01fcf659',container_format='bare',created_at=2024-09-30T16:30:14Z,direct_url=,disk_format='vmdk',id=0d2132dd-40ec-4b44-b509-6f1ee15d96bf,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1020693132-shelved',owner='9b627b25af0c4982a22c8c0bd28641a3',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-09-30T16:30:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1157.341132] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.341132] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1157.341132] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.341132] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1157.341364] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1157.341456] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1157.341621] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1157.342661] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1157.342661] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1157.342661] env[61964]: DEBUG nova.virt.hardware [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1157.343026] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92302cf-f350-42c6-8b1c-ea18dbb2dd89 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.351249] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c9e69e-2eb0-4de3-a238-dbb178365928 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.364829] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:2a:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d7c51d7-90bb-4a31-8ce4-230157e98433', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1157.372708] env[61964]: DEBUG oslo.service.loopingcall [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1157.372972] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1157.373205] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83422c9a-68fc-44d5-a7f0-d97f46ca4e6c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.393748] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1157.393748] env[61964]: value = "task-1041661" [ 1157.393748] env[61964]: _type = "Task" [ 1157.393748] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.401225] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041661, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.635863] env[61964]: DEBUG nova.network.neutron [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.659500] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c30ebb5d-9c58-4033-9cd5-81c019dd13c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.668824] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11121411-2b96-42c2-82fa-b32cb1ecd6e7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.695247] env[61964]: DEBUG nova.compute.manager [req-be1d3af8-d414-4d35-ae93-0b6143800f8c req-72a6d3d0-faa0-4e95-8cc4-bc0c9dbe7be7 service nova] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Detach interface failed, port_id=d7fbf12c-2a5d-4e7f-a51e-b087b2088571, reason: Instance 2da49b28-0cb0-4486-8e69-ce6fba20387c could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1157.903744] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041661, 'name': CreateVM_Task, 'duration_secs': 0.322612} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.903930] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1157.904591] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.904766] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.905163] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1157.905429] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac13d1a-a739-470c-bc4d-f0d5723a7167 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.910357] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1157.910357] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52446857-e8ed-f114-f78a-76028f7a1cea" [ 1157.910357] env[61964]: _type = "Task" [ 1157.910357] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.920198] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52446857-e8ed-f114-f78a-76028f7a1cea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.997817] env[61964]: DEBUG nova.compute.manager [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1157.997954] env[61964]: DEBUG nova.compute.manager [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing instance network info cache due to event network-changed-8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1157.998236] env[61964]: DEBUG oslo_concurrency.lockutils [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] Acquiring lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.998328] env[61964]: DEBUG oslo_concurrency.lockutils [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] Acquired lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.998496] env[61964]: DEBUG nova.network.neutron [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Refreshing network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1158.139240] env[61964]: INFO nova.compute.manager [-] [instance: 2da49b28-0cb0-4486-8e69-ce6fba20387c] Took 1.44 seconds to deallocate network for instance. [ 1158.422996] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.423304] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Processing image 0d2132dd-40ec-4b44-b509-6f1ee15d96bf {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1158.423554] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.423710] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.423890] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1158.424169] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d654109-a1a7-4912-b5c8-96449693d6e8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.439762] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1158.439910] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1158.440607] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6e51783-d922-4115-a919-47dabbd4d22a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.445308] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1158.445308] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52843afc-8fc0-37b2-bb5c-de9f438f137a" [ 1158.445308] env[61964]: _type = "Task" [ 1158.445308] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.452131] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52843afc-8fc0-37b2-bb5c-de9f438f137a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.646028] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.646122] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.646310] env[61964]: DEBUG nova.objects.instance [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'resources' on Instance uuid 2da49b28-0cb0-4486-8e69-ce6fba20387c {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.693040] env[61964]: DEBUG nova.network.neutron [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updated VIF entry in instance network info cache for port 8d7c51d7-90bb-4a31-8ce4-230157e98433. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1158.693407] env[61964]: DEBUG nova.network.neutron [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [{"id": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "address": "fa:16:3e:9c:2a:ce", "network": {"id": "fffa7097-cbb8-4105-8a55-bea2820a0d23", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-977498200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b627b25af0c4982a22c8c0bd28641a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d7c51d7-90", "ovs_interfaceid": "8d7c51d7-90bb-4a31-8ce4-230157e98433", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.955558] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1158.955814] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Fetch image to [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6/OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1158.956020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Downloading stream optimized image 0d2132dd-40ec-4b44-b509-6f1ee15d96bf to [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6/OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6.vmdk on the data store datastore2 as vApp {{(pid=61964) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1158.956208] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Downloading image file data 0d2132dd-40ec-4b44-b509-6f1ee15d96bf to the ESX as VM named 'OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6' {{(pid=61964) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1159.020242] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1159.020242] env[61964]: value = "resgroup-9" [ 1159.020242] env[61964]: _type = "ResourcePool" [ 1159.020242] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1159.020529] env[61964]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c0198de0-3659-4345-a96f-c87d00eda300 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.041605] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease: (returnval){ [ 1159.041605] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521507b1-29f4-5a5e-90ac-5cc77ad76d19" [ 1159.041605] env[61964]: _type = "HttpNfcLease" [ 1159.041605] env[61964]: } obtained for vApp import into resource pool (val){ [ 1159.041605] env[61964]: value = "resgroup-9" [ 1159.041605] env[61964]: _type = "ResourcePool" [ 1159.041605] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1159.041936] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the lease: (returnval){ [ 1159.041936] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521507b1-29f4-5a5e-90ac-5cc77ad76d19" [ 1159.041936] env[61964]: _type = "HttpNfcLease" [ 1159.041936] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1159.047904] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1159.047904] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521507b1-29f4-5a5e-90ac-5cc77ad76d19" [ 1159.047904] env[61964]: _type = "HttpNfcLease" [ 1159.047904] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1159.195386] env[61964]: DEBUG oslo_concurrency.lockutils [req-3b7f8c55-1681-4e62-a87e-f09867153339 req-aff1d7d5-ae04-4b37-973f-9fd9ed036b07 service nova] Releasing lock "refresh_cache-cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.201966] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c07478-a375-4f80-97a7-a002d081c5be {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.210171] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83e2ffc-810d-45bb-b0e1-2fed844c3f2a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.238995] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ceefd47-45a5-43ca-aec5-80e262a6c223 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.245545] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ea74d5-7586-4f1b-8160-e0c8adb5173e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.259841] env[61964]: DEBUG nova.compute.provider_tree [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.550101] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1159.550101] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521507b1-29f4-5a5e-90ac-5cc77ad76d19" [ 1159.550101] env[61964]: _type = "HttpNfcLease" [ 1159.550101] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1159.550563] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1159.550563] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521507b1-29f4-5a5e-90ac-5cc77ad76d19" [ 1159.550563] env[61964]: _type = "HttpNfcLease" [ 1159.550563] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1159.551103] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36ad942-fb7b-4edc-b8e5-974989c00fa9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.557848] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1159.558024] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1159.621071] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-132ea864-2d9d-4449-866c-7965147a119f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.762573] env[61964]: DEBUG nova.scheduler.client.report [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1160.270576] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.291170] env[61964]: INFO nova.scheduler.client.report [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted allocations for instance 2da49b28-0cb0-4486-8e69-ce6fba20387c [ 1160.739537] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1160.739954] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1160.740814] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3060cd5f-385c-4076-8291-ca4092c707ba {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.747603] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1160.747758] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1160.747994] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-74b23b8e-6bc3-4c19-9a5d-5f17df78e92d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.800801] env[61964]: DEBUG oslo_concurrency.lockutils [None req-add5bf63-237b-4d11-b205-e1f34ef2f13b tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "2da49b28-0cb0-4486-8e69-ce6fba20387c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.717s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.069413] env[61964]: DEBUG oslo_vmware.rw_handles [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525196c5-627c-a0c2-5ebd-ad84a66c8bf1/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1161.069643] env[61964]: INFO nova.virt.vmwareapi.images [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Downloaded image file data 0d2132dd-40ec-4b44-b509-6f1ee15d96bf [ 1161.070623] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175b3f81-8030-4420-918e-7e60c8f49b83 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.085624] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1dcc3c39-0473-4d12-ae45-dd8f18f04d73 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.207863] env[61964]: INFO nova.virt.vmwareapi.images [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] The imported VM was unregistered [ 1161.210289] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1161.210539] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Creating directory with path [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1161.210821] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-790b16db-a3d0-4eaf-8e4b-55062fcfb8a0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.229395] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Created directory with path [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1161.229617] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6/OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6.vmdk to [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk. {{(pid=61964) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1161.229888] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-57981cc1-9c81-4706-99d5-470b92bb55af {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.235801] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1161.235801] env[61964]: value = "task-1041664" [ 1161.235801] env[61964]: _type = "Task" [ 1161.235801] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.243179] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.747039] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.247155] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.749147] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.248155] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.326725] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.326968] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.749060] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041664, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.340732} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.749192] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6/OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6.vmdk to [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk. [ 1163.749401] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Cleaning up location [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1163.749537] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e87e99f6-5093-4e0b-8186-2536d3a8e2d6 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.749789] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0201123a-c13d-4c71-946b-bfb9103859e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.755666] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1163.755666] env[61964]: value = "task-1041665" [ 1163.755666] env[61964]: _type = "Task" [ 1163.755666] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.762832] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.832027] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Starting instance... {{(pid=61964) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1164.265976] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033351} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.266253] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.266426] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.266711] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk to [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1164.266966] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b423e77-701d-4151-b8af-7e9e4cb25a37 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.273763] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1164.273763] env[61964]: value = "task-1041666" [ 1164.273763] env[61964]: _type = "Task" [ 1164.273763] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.281112] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.352481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.352746] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.354359] env[61964]: INFO nova.compute.claims [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1164.784340] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.286545] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.421060] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b360409-33a6-42db-824d-6c8e05834b92 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.428646] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac1ae1d-aaa1-4e53-ad69-adf90e51d18a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.461659] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8c79e7-bb95-4ab0-9862-e1a0162c440c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.469875] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0d5698-0939-4fa5-9b69-7ddba68564da {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.484237] env[61964]: DEBUG nova.compute.provider_tree [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.786866] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.987480] env[61964]: DEBUG nova.scheduler.client.report [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1166.286434] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.492727] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.140s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.493307] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Start building networks asynchronously for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1166.787265] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041666, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.190861} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.787464] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0d2132dd-40ec-4b44-b509-6f1ee15d96bf/0d2132dd-40ec-4b44-b509-6f1ee15d96bf.vmdk to [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1166.788105] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced94910-05dd-42b5-815c-248a27686d57 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.808965] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1166.809233] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52e5445e-1720-4ad8-9630-14667d43b93c {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.827568] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1166.827568] env[61964]: value = "task-1041667" [ 1166.827568] env[61964]: _type = "Task" [ 1166.827568] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.834754] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.999030] env[61964]: DEBUG nova.compute.utils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1167.000118] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Allocating IP information in the background. {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1167.000383] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] allocate_for_instance() {{(pid=61964) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1167.039636] env[61964]: DEBUG nova.policy [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d6a5aa43ac34579b8cee43cc6ac9e2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d63f8afe09843efb54de6ea85f276aa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61964) authorize /opt/stack/nova/nova/policy.py:201}} [ 1167.337358] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041667, 'name': ReconfigVM_Task, 'duration_secs': 0.269842} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.337641] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Reconfigured VM instance instance-0000006c to attach disk [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032/cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1167.338269] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5295e1ad-7ebc-4412-ba13-c34cd1fd4065 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.344353] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1167.344353] env[61964]: value = "task-1041668" [ 1167.344353] env[61964]: _type = "Task" [ 1167.344353] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.351770] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041668, 'name': Rename_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.503394] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Start building block device mappings for instance. {{(pid=61964) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1167.545464] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Successfully created port: 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1167.854296] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041668, 'name': Rename_Task, 'duration_secs': 0.145005} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.854668] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1167.854828] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60e1e951-198b-4752-8e22-79a8542301b2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.861231] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1167.861231] env[61964]: value = "task-1041669" [ 1167.861231] env[61964]: _type = "Task" [ 1167.861231] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.868237] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041669, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.370973] env[61964]: DEBUG oslo_vmware.api [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041669, 'name': PowerOnVM_Task, 'duration_secs': 0.441222} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.371327] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1168.479275] env[61964]: DEBUG nova.compute.manager [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1168.480195] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fc4ec1-30aa-4b30-ac04-46acf74a01cc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.513086] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Start spawning the instance on the hypervisor. {{(pid=61964) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1168.533062] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T16:18:12Z,direct_url=,disk_format='vmdk',id=5f28385d-6ea0-420d-8a26-4cb693714c14,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c5db75cf61c441d396a6af209d5f2d11',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T16:18:13Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1168.533348] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1168.533559] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1168.533771] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1168.533932] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1168.534093] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1168.534302] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1168.534699] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1168.534894] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1168.535075] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1168.535255] env[61964]: DEBUG nova.virt.hardware [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1168.536209] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d477f7ad-564b-4a7f-842e-46c047b3bf28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.544316] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786b2ed2-17b7-4dc9-b72a-80c269588fc7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.918054] env[61964]: DEBUG nova.compute.manager [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1168.918054] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.918054] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.918054] env[61964]: DEBUG oslo_concurrency.lockutils [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.918054] env[61964]: DEBUG nova.compute.manager [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] No waiting events found dispatching network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1168.918054] env[61964]: WARNING nova.compute.manager [req-1c35c0d9-4a91-4ea0-afdf-85a5837f4433 req-49a6a67f-97bb-4a27-a430-3c42d6f192f8 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received unexpected event network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 for instance with vm_state building and task_state spawning. [ 1168.998665] env[61964]: DEBUG oslo_concurrency.lockutils [None req-12e41a44-f2ce-471a-bab6-5b1757652804 tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.701s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.005458] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Successfully updated port: 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1169.507648] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1169.507813] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.507971] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1170.039457] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance cache missing network info. {{(pid=61964) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1170.155487] env[61964]: DEBUG nova.network.neutron [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.315303] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.315613] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.315836] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.316042] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.316221] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.318397] env[61964]: INFO nova.compute.manager [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Terminating instance [ 1170.658288] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.658595] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance network_info: |[{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61964) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1170.659045] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:20:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c776c76-6266-46eb-8acb-3896269c2dc5', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1170.666482] env[61964]: DEBUG oslo.service.loopingcall [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1170.666704] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1170.666952] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cbbb5ba-7611-41af-95b6-5e46825f658f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.686840] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1170.686840] env[61964]: value = "task-1041670" [ 1170.686840] env[61964]: _type = "Task" [ 1170.686840] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.694396] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041670, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.822651] env[61964]: DEBUG nova.compute.manager [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1170.822891] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1170.823876] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da248b7-7bd4-4327-b390-a00195c1d918 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.831601] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1170.831859] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58824643-abdd-42d5-97ec-38b51a6976c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.838067] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1170.838067] env[61964]: value = "task-1041671" [ 1170.838067] env[61964]: _type = "Task" [ 1170.838067] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.848636] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.949371] env[61964]: DEBUG nova.compute.manager [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1170.949584] env[61964]: DEBUG nova.compute.manager [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing instance network info cache due to event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1170.949807] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.949966] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.950150] env[61964]: DEBUG nova.network.neutron [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1171.196792] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041670, 'name': CreateVM_Task, 'duration_secs': 0.313559} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.197164] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1171.197712] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.197920] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.198305] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1171.198597] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84d20c1-d20b-4423-99c9-2663aac5484e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.202845] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1171.202845] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52357992-43aa-a03a-49b8-453200f32b0e" [ 1171.202845] env[61964]: _type = "Task" [ 1171.202845] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.210731] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52357992-43aa-a03a-49b8-453200f32b0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.347285] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041671, 'name': PowerOffVM_Task, 'duration_secs': 0.207045} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.347554] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1171.347733] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1171.348000] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a69a69d-a47e-4bed-9572-63454574494d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.408359] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1171.408586] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1171.408774] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleting the datastore file [datastore2] cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1171.409064] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22a80810-e7f0-4da3-8ac3-0dc6f842cab1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.415227] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for the task: (returnval){ [ 1171.415227] env[61964]: value = "task-1041673" [ 1171.415227] env[61964]: _type = "Task" [ 1171.415227] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.422325] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.650252] env[61964]: DEBUG nova.network.neutron [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated VIF entry in instance network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1171.650645] env[61964]: DEBUG nova.network.neutron [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.712381] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52357992-43aa-a03a-49b8-453200f32b0e, 'name': SearchDatastore_Task, 'duration_secs': 0.010942} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.712635] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.712870] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Processing image 5f28385d-6ea0-420d-8a26-4cb693714c14 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1171.713123] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.713288] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.713513] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1171.713769] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56e9cd8b-6176-43f9-9e33-c453d8bffe78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.721645] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1171.721819] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1171.722516] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-247a58bf-614f-4a1f-9297-7d2b0f4c0c47 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.727177] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1171.727177] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d7340-5c5c-1aae-1ebe-4c47c3ee1ff1" [ 1171.727177] env[61964]: _type = "Task" [ 1171.727177] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.734231] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d7340-5c5c-1aae-1ebe-4c47c3ee1ff1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.925164] env[61964]: DEBUG oslo_vmware.api [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Task: {'id': task-1041673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140045} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.925424] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.925640] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.925835] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.926029] env[61964]: INFO nova.compute.manager [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1171.926277] env[61964]: DEBUG oslo.service.loopingcall [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.926492] env[61964]: DEBUG nova.compute.manager [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1171.926593] env[61964]: DEBUG nova.network.neutron [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1172.153185] env[61964]: DEBUG oslo_concurrency.lockutils [req-1a95f303-9916-45dd-a4d6-66f4a8ace2bb req-f6774c29-ff51-4c22-988f-13c8fb6e3375 service nova] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.237790] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]522d7340-5c5c-1aae-1ebe-4c47c3ee1ff1, 'name': SearchDatastore_Task, 'duration_secs': 0.009554} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.238579] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdf6d35b-f1c6-431c-9c95-979c8decbf65 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.243427] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1172.243427] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280c1de-01f4-52dd-f166-ea1dc278199f" [ 1172.243427] env[61964]: _type = "Task" [ 1172.243427] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.251568] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280c1de-01f4-52dd-f166-ea1dc278199f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.754047] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]5280c1de-01f4-52dd-f166-ea1dc278199f, 'name': SearchDatastore_Task, 'duration_secs': 0.009631} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.754282] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.754542] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1172.754801] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-210fd3a2-ac53-4832-9b93-3dd549a0a4bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.760754] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1172.760754] env[61964]: value = "task-1041674" [ 1172.760754] env[61964]: _type = "Task" [ 1172.760754] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.768236] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041674, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.862852] env[61964]: DEBUG nova.network.neutron [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.978231] env[61964]: DEBUG nova.compute.manager [req-0e4906d7-9518-4b74-8722-9a4bbfdc6e11 req-cee18394-3d2a-4306-bb69-eff04de477cd service nova] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Received event network-vif-deleted-8d7c51d7-90bb-4a31-8ce4-230157e98433 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1173.270723] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041674, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467267} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.272020] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f28385d-6ea0-420d-8a26-4cb693714c14/5f28385d-6ea0-420d-8a26-4cb693714c14.vmdk to [datastore1] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1173.272020] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Extending root virtual disk to 1048576 {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1173.272020] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b268ab1-faf8-4b6c-896d-44fac43de0ee {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.278817] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1173.278817] env[61964]: value = "task-1041675" [ 1173.278817] env[61964]: _type = "Task" [ 1173.278817] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.286358] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041675, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.366209] env[61964]: INFO nova.compute.manager [-] [instance: cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032] Took 1.44 seconds to deallocate network for instance. [ 1173.789669] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041675, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066496} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.790065] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Extended root virtual disk {{(pid=61964) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1173.790832] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20366e5c-1640-4c75-8b5c-dd225cbea230 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.811896] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.812178] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f563498-f3f9-488d-beed-d1f902d2d22e {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.830658] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1173.830658] env[61964]: value = "task-1041676" [ 1173.830658] env[61964]: _type = "Task" [ 1173.830658] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.840476] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.872595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.872816] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.873058] env[61964]: DEBUG nova.objects.instance [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lazy-loading 'resources' on Instance uuid cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.340283] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041676, 'name': ReconfigVM_Task, 'duration_secs': 0.279515} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.340605] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Reconfigured VM instance instance-0000006e to attach disk [datastore1] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk or device None with type sparse {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.341267] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00ad6894-3d79-414e-9227-a401189fd4e4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.346759] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1174.346759] env[61964]: value = "task-1041677" [ 1174.346759] env[61964]: _type = "Task" [ 1174.346759] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.358227] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041677, 'name': Rename_Task} progress is 6%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.427387] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e4da78-c021-4a87-9a4a-4bb8fcbff1b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.433912] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011806e5-9af1-4604-9f3b-f52be79cf614 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.464623] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608511bb-3775-4834-ac77-3a84cad3e4f7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.472178] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a137d97e-a94e-46e6-aeb3-d7067201d52b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.486413] env[61964]: DEBUG nova.compute.provider_tree [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.856784] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041677, 'name': Rename_Task, 'duration_secs': 0.128499} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.857087] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1174.857338] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5859b4ff-be2b-48e4-a2b4-7a3ebc84e274 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.863193] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1174.863193] env[61964]: value = "task-1041678" [ 1174.863193] env[61964]: _type = "Task" [ 1174.863193] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.869961] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041678, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.992631] env[61964]: DEBUG nova.scheduler.client.report [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.372671] env[61964]: DEBUG oslo_vmware.api [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041678, 'name': PowerOnVM_Task, 'duration_secs': 0.41465} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.373084] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1175.373161] env[61964]: INFO nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Took 6.86 seconds to spawn the instance on the hypervisor. [ 1175.373413] env[61964]: DEBUG nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1175.374199] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79748235-69d5-4229-a6b4-8487431970a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.498687] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.515888] env[61964]: INFO nova.scheduler.client.report [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Deleted allocations for instance cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032 [ 1175.891322] env[61964]: INFO nova.compute.manager [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Took 11.56 seconds to build instance. [ 1176.023161] env[61964]: DEBUG oslo_concurrency.lockutils [None req-64fe11d8-e7e4-425a-b877-9d41648cd2ab tempest-ServerActionsTestOtherB-1406292670 tempest-ServerActionsTestOtherB-1406292670-project-member] Lock "cb1174c3-eb5f-43cb-a1c6-6d31c9d3a032" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.707s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.394052] env[61964]: DEBUG oslo_concurrency.lockutils [None req-c4871081-f8f8-4c63-badf-58d2daecbb3e tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.067s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.471722] env[61964]: DEBUG nova.compute.manager [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1176.471868] env[61964]: DEBUG nova.compute.manager [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing instance network info cache due to event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1176.472106] env[61964]: DEBUG oslo_concurrency.lockutils [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.472259] env[61964]: DEBUG oslo_concurrency.lockutils [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.472425] env[61964]: DEBUG nova.network.neutron [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1177.183394] env[61964]: DEBUG nova.network.neutron [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated VIF entry in instance network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1177.183770] env[61964]: DEBUG nova.network.neutron [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.686752] env[61964]: DEBUG oslo_concurrency.lockutils [req-361e56f6-2f4d-4c6c-a9f5-ece20433c1de req-1d7ede6b-cf14-4719-bf4e-23a9333a9e05 service nova] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.785866] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.786140] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.289100] env[61964]: DEBUG nova.compute.utils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Using /dev/sd instead of None {{(pid=61964) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1180.792688] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.857516] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.857915] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.858073] env[61964]: INFO nova.compute.manager [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Attaching volume ea264dbd-203f-4e32-aa2e-e5b1e8769ffb to /dev/sdb [ 1181.888087] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75507d0c-e9d0-4722-8bdc-a488baa40981 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.895266] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2997f38-1725-478f-95ad-4ad64d5c5cdd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.907755] env[61964]: DEBUG nova.virt.block_device [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating existing volume attachment record: 80767967-588b-4654-876d-93967700c6e9 {{(pid=61964) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1186.450562] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Volume attach. Driver type: vmdk {{(pid=61964) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1186.450842] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230588', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'name': 'volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '85355640-f271-4e0a-bf75-0bbcb8487d6f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'serial': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1186.451727] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4d16bf-6060-401d-8986-670fb2cff37b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.467682] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8a03ce-2bfe-461c-9286-2f4e56e127b0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.491014] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb/volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.491258] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2308e47a-790e-4040-a6bf-82458a519947 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.508170] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1186.508170] env[61964]: value = "task-1041682" [ 1186.508170] env[61964]: _type = "Task" [ 1186.508170] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.515461] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041682, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.018025] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041682, 'name': ReconfigVM_Task, 'duration_secs': 0.313607} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.018338] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb/volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb.vmdk or device None with type thin {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1187.022931] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96835506-1774-4c5a-822b-cbb021bb3486 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.036897] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1187.036897] env[61964]: value = "task-1041683" [ 1187.036897] env[61964]: _type = "Task" [ 1187.036897] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.044453] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041683, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.546141] env[61964]: DEBUG oslo_vmware.api [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041683, 'name': ReconfigVM_Task, 'duration_secs': 0.131028} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.546528] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230588', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'name': 'volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '85355640-f271-4e0a-bf75-0bbcb8487d6f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'serial': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb'} {{(pid=61964) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1188.580548] env[61964]: DEBUG nova.objects.instance [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 85355640-f271-4e0a-bf75-0bbcb8487d6f {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.086639] env[61964]: DEBUG oslo_concurrency.lockutils [None req-3d428636-d8d6-445d-be2b-0e082ba8f95e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.229s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.265809] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.266188] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.769760] env[61964]: INFO nova.compute.manager [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Detaching volume ea264dbd-203f-4e32-aa2e-e5b1e8769ffb [ 1189.798962] env[61964]: INFO nova.virt.block_device [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Attempting to driver detach volume ea264dbd-203f-4e32-aa2e-e5b1e8769ffb from mountpoint /dev/sdb [ 1189.799222] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Volume detach. Driver type: vmdk {{(pid=61964) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1189.799416] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230588', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'name': 'volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '85355640-f271-4e0a-bf75-0bbcb8487d6f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'serial': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1189.800305] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a150046-4328-435f-b4ee-9f1adf67f21f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.821748] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd1db31-8e8f-4287-9acd-06a09172c185 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.828323] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343e077c-1d1d-4c11-9546-847379f1f80f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.847437] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302554e8-6104-4b1c-9f10-219cbb686ff7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.862177] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] The volume has not been displaced from its original location: [datastore2] volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb/volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb.vmdk. No consolidation needed. {{(pid=61964) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1189.867284] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1189.867542] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c9ba4f5-b947-4cbe-a896-251489a72558 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.884442] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1189.884442] env[61964]: value = "task-1041684" [ 1189.884442] env[61964]: _type = "Task" [ 1189.884442] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.891759] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.393942] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041684, 'name': ReconfigVM_Task, 'duration_secs': 0.221428} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.394241] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=61964) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1190.398807] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c73d2016-d530-4418-ab43-098e03f22fa5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.413202] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1190.413202] env[61964]: value = "task-1041685" [ 1190.413202] env[61964]: _type = "Task" [ 1190.413202] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.420991] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.923873] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041685, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.424133] env[61964]: DEBUG oslo_vmware.api [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041685, 'name': ReconfigVM_Task, 'duration_secs': 0.76241} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.424450] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-230588', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'name': 'volume-ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '85355640-f271-4e0a-bf75-0bbcb8487d6f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb', 'serial': 'ea264dbd-203f-4e32-aa2e-e5b1e8769ffb'} {{(pid=61964) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1191.964895] env[61964]: DEBUG nova.objects.instance [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'flavor' on Instance uuid 85355640-f271-4e0a-bf75-0bbcb8487d6f {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.971975] env[61964]: DEBUG oslo_concurrency.lockutils [None req-a9a29768-5ae8-4fed-951a-6e357cc7a88e tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.706s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.537326] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.537700] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.537940] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.538164] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.538338] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.540681] env[61964]: INFO nova.compute.manager [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Terminating instance [ 1194.044920] env[61964]: DEBUG nova.compute.manager [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1194.045428] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1194.046167] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8e392f-e67b-4149-844e-85f062f330bc {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.054204] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1194.054460] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c416585-8755-4314-8931-d6a895b027b6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.061333] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1194.061333] env[61964]: value = "task-1041686" [ 1194.061333] env[61964]: _type = "Task" [ 1194.061333] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.070366] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.571078] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041686, 'name': PowerOffVM_Task, 'duration_secs': 0.162451} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.571391] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1194.571569] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1194.571814] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-488dd0c9-72ae-466c-87d8-26de076777e6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.636733] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1194.636915] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1194.637136] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleting the datastore file [datastore1] 85355640-f271-4e0a-bf75-0bbcb8487d6f {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1194.637419] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21b46ec0-2e5a-41c7-bd6f-8f013a2b05e1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.645945] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for the task: (returnval){ [ 1194.645945] env[61964]: value = "task-1041688" [ 1194.645945] env[61964]: _type = "Task" [ 1194.645945] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.653016] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.156871] env[61964]: DEBUG oslo_vmware.api [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Task: {'id': task-1041688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136585} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.157274] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1195.157374] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1195.157515] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1195.157697] env[61964]: INFO nova.compute.manager [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1195.157938] env[61964]: DEBUG oslo.service.loopingcall [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1195.158147] env[61964]: DEBUG nova.compute.manager [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1195.158241] env[61964]: DEBUG nova.network.neutron [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1195.584155] env[61964]: DEBUG nova.compute.manager [req-4b765926-e74c-4ab0-b21d-df688532c0e6 req-c0ef9ad2-ce23-4d18-8a8a-80fd3c748c19 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Received event network-vif-deleted-6f4ceb27-b4c3-4358-8596-ac8c27674bdf {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1195.584390] env[61964]: INFO nova.compute.manager [req-4b765926-e74c-4ab0-b21d-df688532c0e6 req-c0ef9ad2-ce23-4d18-8a8a-80fd3c748c19 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Neutron deleted interface 6f4ceb27-b4c3-4358-8596-ac8c27674bdf; detaching it from the instance and deleting it from the info cache [ 1195.584550] env[61964]: DEBUG nova.network.neutron [req-4b765926-e74c-4ab0-b21d-df688532c0e6 req-c0ef9ad2-ce23-4d18-8a8a-80fd3c748c19 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.064011] env[61964]: DEBUG nova.network.neutron [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.086664] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43880b41-43e9-47e3-b67a-759b02b798a9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.097093] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c75eb4e-3b52-430a-aacb-205e7ac196b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.121512] env[61964]: DEBUG nova.compute.manager [req-4b765926-e74c-4ab0-b21d-df688532c0e6 req-c0ef9ad2-ce23-4d18-8a8a-80fd3c748c19 service nova] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Detach interface failed, port_id=6f4ceb27-b4c3-4358-8596-ac8c27674bdf, reason: Instance 85355640-f271-4e0a-bf75-0bbcb8487d6f could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1196.569057] env[61964]: INFO nova.compute.manager [-] [instance: 85355640-f271-4e0a-bf75-0bbcb8487d6f] Took 1.41 seconds to deallocate network for instance. [ 1197.073781] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.074079] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.074329] env[61964]: DEBUG nova.objects.instance [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lazy-loading 'resources' on Instance uuid 85355640-f271-4e0a-bf75-0bbcb8487d6f {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.619051] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b8f65a-295e-45e1-912d-1b8105037d7d {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.626212] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e12faa9-2843-4b9f-b8b4-3808ae9cfbc5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.656366] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40da722-4cd5-456c-b143-058e0f94a830 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.663675] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027650b3-683f-472c-9c90-c25549a0e6c4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.676537] env[61964]: DEBUG nova.compute.provider_tree [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.180060] env[61964]: DEBUG nova.scheduler.client.report [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1198.684067] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.703256] env[61964]: INFO nova.scheduler.client.report [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Deleted allocations for instance 85355640-f271-4e0a-bf75-0bbcb8487d6f [ 1199.212218] env[61964]: DEBUG oslo_concurrency.lockutils [None req-ef438824-8cf2-48af-be7f-0ffb2a75cc00 tempest-AttachVolumeNegativeTest-1485752094 tempest-AttachVolumeNegativeTest-1485752094-project-member] Lock "85355640-f271-4e0a-bf75-0bbcb8487d6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.674s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.473422] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.473640] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.473766] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1199.473887] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 1200.004550] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.004822] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.004883] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1200.005018] env[61964]: DEBUG nova.objects.instance [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lazy-loading 'info_cache' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.709380] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.211691] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.211928] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 1202.212150] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.212324] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.212511] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.212684] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.212832] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.212981] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.213139] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1202.213280] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.716818] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.717194] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.717194] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.717362] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1202.718247] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d210cfd0-13e0-46ab-b0f8-29a1a5e15fca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.726389] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2c6cd9-80a8-41a2-850a-26bd90f8c519 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.740193] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3e74e5-a38c-4458-801f-0f86cb494632 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.746141] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d196f8-7a74-4cac-ab97-09e8c4265917 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.773593] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181334MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1202.773742] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.773908] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.796899] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a5ee87bf-770f-45be-83d5-16d175102430 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.797189] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1203.797273] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1203.821161] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4588cf9f-2de4-48f5-8788-08a33cf35999 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.828422] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0884cd2-91a9-426c-b400-211a2d528f0f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.858228] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e5422b-564e-4014-89dd-d7366d2d0066 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.864823] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c13c94-4b3a-4765-8aed-94d08ff83e05 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.877179] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1204.380448] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.885994] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1204.886447] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.112s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.173685] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1215.627595] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.627963] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.628138] env[61964]: INFO nova.compute.manager [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Shelving [ 1216.638610] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1216.638979] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b14619ab-9be3-4d72-855f-0e391aaca8d5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.645757] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1216.645757] env[61964]: value = "task-1041692" [ 1216.645757] env[61964]: _type = "Task" [ 1216.645757] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.653465] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.156412] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041692, 'name': PowerOffVM_Task, 'duration_secs': 0.178418} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.156699] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1217.157477] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d026c8eb-08a0-4bce-93d4-aa3e923424ca {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.175185] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2d296c-2c74-4044-9ee2-8f3b0160db31 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.685543] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Creating Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1217.685984] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-85f4000c-b747-45b1-9a79-4ff03de22546 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.694251] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1217.694251] env[61964]: value = "task-1041693" [ 1217.694251] env[61964]: _type = "Task" [ 1217.694251] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.702698] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041693, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.204799] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041693, 'name': CreateSnapshot_Task, 'duration_secs': 0.396089} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.205079] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Created Snapshot of the VM instance {{(pid=61964) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1218.205855] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc52b23c-e18e-4ca5-8869-8469018a81ce {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.723301] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Creating linked-clone VM from snapshot {{(pid=61964) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1218.723688] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2ea6e4c0-9a10-44aa-90a6-c1b8a320a815 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.732344] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1218.732344] env[61964]: value = "task-1041694" [ 1218.732344] env[61964]: _type = "Task" [ 1218.732344] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.741290] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041694, 'name': CloneVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.242161] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041694, 'name': CloneVM_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.742325] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041694, 'name': CloneVM_Task, 'duration_secs': 0.880989} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.742655] env[61964]: INFO nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Created linked-clone VM from snapshot [ 1219.743411] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c142723-62f3-4530-8fae-9861be84842f {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.750395] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Uploading image eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1219.770679] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1219.770679] env[61964]: value = "vm-230590" [ 1219.770679] env[61964]: _type = "VirtualMachine" [ 1219.770679] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1219.770933] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4054237d-429b-4220-a2d2-0ad96a15b937 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.777786] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease: (returnval){ [ 1219.777786] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52025daf-ef73-5f3b-44dc-47bd62c1f003" [ 1219.777786] env[61964]: _type = "HttpNfcLease" [ 1219.777786] env[61964]: } obtained for exporting VM: (result){ [ 1219.777786] env[61964]: value = "vm-230590" [ 1219.777786] env[61964]: _type = "VirtualMachine" [ 1219.777786] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1219.778088] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the lease: (returnval){ [ 1219.778088] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52025daf-ef73-5f3b-44dc-47bd62c1f003" [ 1219.778088] env[61964]: _type = "HttpNfcLease" [ 1219.778088] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1219.783550] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1219.783550] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52025daf-ef73-5f3b-44dc-47bd62c1f003" [ 1219.783550] env[61964]: _type = "HttpNfcLease" [ 1219.783550] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1220.288149] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1220.288149] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52025daf-ef73-5f3b-44dc-47bd62c1f003" [ 1220.288149] env[61964]: _type = "HttpNfcLease" [ 1220.288149] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1220.288485] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1220.288485] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52025daf-ef73-5f3b-44dc-47bd62c1f003" [ 1220.288485] env[61964]: _type = "HttpNfcLease" [ 1220.288485] env[61964]: }. {{(pid=61964) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1220.289211] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8281e223-ab52-4ee6-ba0e-1d6d03a23144 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.296032] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1220.296211] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk for reading. {{(pid=61964) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1220.382882] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6c8c7463-badf-477e-a19a-a5e502e4e2c8 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.316688] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1228.317710] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c343a0-97a8-4109-8a17-45268f66e0ab {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.323954] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1228.324150] env[61964]: ERROR oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk due to incomplete transfer. [ 1228.324374] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-15b4f227-0dd2-4fe6-a11f-08d55b2cdce2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.331863] env[61964]: DEBUG oslo_vmware.rw_handles [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fe07c2-9cf3-2a87-eb0b-d1bca4718888/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1228.332079] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Uploaded image eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e to the Glance image server {{(pid=61964) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1228.334530] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Destroying the VM {{(pid=61964) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1228.334767] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-26d2e90f-2d7f-4bee-9833-94e52329f006 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.340767] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1228.340767] env[61964]: value = "task-1041696" [ 1228.340767] env[61964]: _type = "Task" [ 1228.340767] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.348864] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041696, 'name': Destroy_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.850554] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041696, 'name': Destroy_Task, 'duration_secs': 0.346636} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.850767] env[61964]: INFO nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Destroyed the VM [ 1228.851012] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleting Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1228.851292] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6a7f461f-a513-452e-8a39-88f279c4f49a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.857901] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1228.857901] env[61964]: value = "task-1041697" [ 1228.857901] env[61964]: _type = "Task" [ 1228.857901] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.865086] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041697, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.367864] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041697, 'name': RemoveSnapshot_Task, 'duration_secs': 0.321743} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.368317] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleted Snapshot of the VM instance {{(pid=61964) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1229.368428] env[61964]: DEBUG nova.compute.manager [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1229.369179] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eba21b5-033f-4d00-ba73-93f477b90575 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.880857] env[61964]: INFO nova.compute.manager [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Shelve offloading [ 1230.384740] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1230.385112] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f68496e-3d1f-4ff2-836d-76c3bc94385a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.392402] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1230.392402] env[61964]: value = "task-1041698" [ 1230.392402] env[61964]: _type = "Task" [ 1230.392402] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.400015] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.903570] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] VM already powered off {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1230.903818] env[61964]: DEBUG nova.compute.manager [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1230.904585] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2127037-465c-4e3a-8537-929667163708 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.910073] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.910247] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.910420] env[61964]: DEBUG nova.network.neutron [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1231.672367] env[61964]: DEBUG nova.network.neutron [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.175524] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.388299] env[61964]: DEBUG nova.compute.manager [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-vif-unplugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1232.388547] env[61964]: DEBUG oslo_concurrency.lockutils [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.388762] env[61964]: DEBUG oslo_concurrency.lockutils [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.388939] env[61964]: DEBUG oslo_concurrency.lockutils [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.389131] env[61964]: DEBUG nova.compute.manager [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] No waiting events found dispatching network-vif-unplugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1232.389313] env[61964]: WARNING nova.compute.manager [req-33e95bb3-2db0-43be-86f6-4ba46b6c5d1b req-24146536-135b-40a4-b6d8-80e808f7bace service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received unexpected event network-vif-unplugged-4c776c76-6266-46eb-8acb-3896269c2dc5 for instance with vm_state shelved and task_state shelving_offloading. [ 1232.465691] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1232.466704] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3192d54e-11ce-4721-9c75-8b7b848f517b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.474119] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1232.474347] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9308bd34-aa96-4da9-bb50-3dddbcceacf6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.542340] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1232.542592] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleting contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1232.542758] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore1] a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.543025] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5180cd52-450a-4a23-bec5-c6b7ef5ee7b3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.549210] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1232.549210] env[61964]: value = "task-1041700" [ 1232.549210] env[61964]: _type = "Task" [ 1232.549210] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.556444] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.058631] env[61964]: DEBUG oslo_vmware.api [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134381} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.059014] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1233.059063] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleted contents of the VM from datastore datastore1 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1233.059250] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1233.087127] env[61964]: INFO nova.scheduler.client.report [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted allocations for instance a5ee87bf-770f-45be-83d5-16d175102430 [ 1233.591667] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.591960] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.592212] env[61964]: DEBUG nova.objects.instance [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'resources' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.095687] env[61964]: DEBUG nova.objects.instance [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'numa_topology' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.412269] env[61964]: DEBUG nova.compute.manager [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1234.412467] env[61964]: DEBUG nova.compute.manager [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing instance network info cache due to event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1234.412692] env[61964]: DEBUG oslo_concurrency.lockutils [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.412842] env[61964]: DEBUG oslo_concurrency.lockutils [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.413015] env[61964]: DEBUG nova.network.neutron [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1234.599263] env[61964]: DEBUG nova.objects.base [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61964) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1234.612761] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431c118d-22d3-4968-8615-249fd6f5ab9a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.620495] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74125c87-f45c-4f25-acc6-d3c9541fecb7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.650740] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7869b678-a2f5-4386-9a1f-c83fef2a1fd9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.657812] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651feb02-1dd2-428c-b812-5468cf484452 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.670237] env[61964]: DEBUG nova.compute.provider_tree [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.129513] env[61964]: DEBUG nova.network.neutron [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated VIF entry in instance network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1235.129898] env[61964]: DEBUG nova.network.neutron [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4c776c76-62", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.172757] env[61964]: DEBUG nova.scheduler.client.report [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1235.633817] env[61964]: DEBUG oslo_concurrency.lockutils [req-4344121d-600d-45d7-be6b-06057c3945ed req-48f4f806-99a5-448e-bc86-0650a3b619f6 service nova] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.677900] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.185920] env[61964]: DEBUG oslo_concurrency.lockutils [None req-d150fdcb-35da-4888-9ed0-b9e0f9d20da9 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.558s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.450368] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1236.450662] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1236.450857] env[61964]: INFO nova.compute.manager [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Unshelving [ 1237.475093] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.475417] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.475653] env[61964]: DEBUG nova.objects.instance [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'pci_requests' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.979675] env[61964]: DEBUG nova.objects.instance [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'numa_topology' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1238.482574] env[61964]: INFO nova.compute.claims [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1239.515961] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d015ce8a-f14f-4181-92ad-f86457732924 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.523304] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f31765d-1358-4f45-810b-edca77d9d0d3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.552094] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f981bdde-ce82-4225-8678-b0244d2b1c78 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.558797] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231d0ec3-bcaa-488d-bd0c-0fa7a61467a5 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.571279] env[61964]: DEBUG nova.compute.provider_tree [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.074577] env[61964]: DEBUG nova.scheduler.client.report [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1240.580146] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.105s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.608457] env[61964]: INFO nova.network.neutron [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating port 4c776c76-6266-46eb-8acb-3896269c2dc5 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1242.012032] env[61964]: DEBUG nova.compute.manager [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1242.012295] env[61964]: DEBUG oslo_concurrency.lockutils [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1242.012501] env[61964]: DEBUG oslo_concurrency.lockutils [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1242.012677] env[61964]: DEBUG oslo_concurrency.lockutils [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.012848] env[61964]: DEBUG nova.compute.manager [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] No waiting events found dispatching network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1242.013280] env[61964]: WARNING nova.compute.manager [req-a6e7cd48-e5db-4e47-8df9-6a2396712f8f req-c36b006b-28ee-4ed7-a5f1-23852a3c780a service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received unexpected event network-vif-plugged-4c776c76-6266-46eb-8acb-3896269c2dc5 for instance with vm_state shelved_offloaded and task_state spawning. [ 1242.301944] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1242.301944] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.302194] env[61964]: DEBUG nova.network.neutron [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Building network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1243.001348] env[61964]: DEBUG nova.network.neutron [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.503984] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1243.530600] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T16:18:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7fba92b7a8833161c36a9dd723576537',container_format='bare',created_at=2024-09-30T16:31:43Z,direct_url=,disk_format='vmdk',id=eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1107306957-shelved',owner='4d63f8afe09843efb54de6ea85f276aa',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-09-30T16:31:57Z,virtual_size=,visibility=), allow threads: False {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1243.530870] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1243.531045] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image limits 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1243.531244] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Flavor pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1243.531399] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Image pref 0:0:0 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1243.531556] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61964) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1243.531762] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1243.531927] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1243.532115] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Got 1 possible topologies {{(pid=61964) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1243.532285] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1243.532464] env[61964]: DEBUG nova.virt.hardware [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61964) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1243.533629] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3a6385-92c4-4d7b-8454-733bbcbb6125 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.541568] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b624bd-0730-43fa-aa31-59e2a107ade0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.554820] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:20:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f3d7e184-c87f-47a5-8d0d-9fa20e07e669', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c776c76-6266-46eb-8acb-3896269c2dc5', 'vif_model': 'vmxnet3'}] {{(pid=61964) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.562049] env[61964]: DEBUG oslo.service.loopingcall [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.562282] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Creating VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1243.562480] env[61964]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e3c7670-3996-44f3-863f-a53754300144 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.581360] env[61964]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.581360] env[61964]: value = "task-1041701" [ 1243.581360] env[61964]: _type = "Task" [ 1243.581360] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.588316] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041701, 'name': CreateVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.090889] env[61964]: DEBUG oslo_vmware.api [-] Task: {'id': task-1041701, 'name': CreateVM_Task, 'duration_secs': 0.314548} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.091060] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Created VM on the ESX host {{(pid=61964) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1244.091696] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.091871] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.092284] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1244.092545] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc611ae1-c7d3-43f4-947e-154381907d79 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.097326] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1244.097326] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527d105e-33a5-d480-cedc-2c6bd4b18f73" [ 1244.097326] env[61964]: _type = "Task" [ 1244.097326] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.104398] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]527d105e-33a5-d480-cedc-2c6bd4b18f73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.217023] env[61964]: DEBUG nova.compute.manager [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1244.217237] env[61964]: DEBUG nova.compute.manager [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing instance network info cache due to event network-changed-4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11578}} [ 1244.217462] env[61964]: DEBUG oslo_concurrency.lockutils [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.217608] env[61964]: DEBUG oslo_concurrency.lockutils [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.217810] env[61964]: DEBUG nova.network.neutron [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Refreshing network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1244.607697] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.608117] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Processing image eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.608204] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.608358] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquired lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.608572] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.608817] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87f2e262-fb69-4cc9-a9dd-de4bb7df1e77 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.616820] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.616961] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61964) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1244.617613] env[61964]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ca30636-188f-4764-85c3-f046e012ac91 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.622447] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1244.622447] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521debc7-44bb-555f-f74e-84c0e8c1f634" [ 1244.622447] env[61964]: _type = "Task" [ 1244.622447] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.630725] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]521debc7-44bb-555f-f74e-84c0e8c1f634, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.920216] env[61964]: DEBUG nova.network.neutron [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated VIF entry in instance network info cache for port 4c776c76-6266-46eb-8acb-3896269c2dc5. {{(pid=61964) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1244.920603] env[61964]: DEBUG nova.network.neutron [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.131648] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Preparing fetch location {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1245.131926] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Fetch image to [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219/OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219.vmdk {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1245.132138] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Downloading stream optimized image eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e to [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219/OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219.vmdk on the data store datastore2 as vApp {{(pid=61964) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1245.132319] env[61964]: DEBUG nova.virt.vmwareapi.images [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Downloading image file data eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e to the ESX as VM named 'OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219' {{(pid=61964) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1245.196031] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1245.196031] env[61964]: value = "resgroup-9" [ 1245.196031] env[61964]: _type = "ResourcePool" [ 1245.196031] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1245.196293] env[61964]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-03c715ce-406e-4491-b005-ed754cdae42a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.216182] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease: (returnval){ [ 1245.216182] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a27453-9094-cb89-1b30-d2ce1a4841d6" [ 1245.216182] env[61964]: _type = "HttpNfcLease" [ 1245.216182] env[61964]: } obtained for vApp import into resource pool (val){ [ 1245.216182] env[61964]: value = "resgroup-9" [ 1245.216182] env[61964]: _type = "ResourcePool" [ 1245.216182] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1245.216639] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the lease: (returnval){ [ 1245.216639] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a27453-9094-cb89-1b30-d2ce1a4841d6" [ 1245.216639] env[61964]: _type = "HttpNfcLease" [ 1245.216639] env[61964]: } to be ready. {{(pid=61964) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1245.222301] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1245.222301] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a27453-9094-cb89-1b30-d2ce1a4841d6" [ 1245.222301] env[61964]: _type = "HttpNfcLease" [ 1245.222301] env[61964]: } is initializing. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1245.424134] env[61964]: DEBUG oslo_concurrency.lockutils [req-00ec4d9f-fc82-4e14-bed1-4b62ed0c7b00 req-501f88e8-bbff-442b-ba8c-51b865ecda6f service nova] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.724889] env[61964]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1245.724889] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a27453-9094-cb89-1b30-d2ce1a4841d6" [ 1245.724889] env[61964]: _type = "HttpNfcLease" [ 1245.724889] env[61964]: } is ready. {{(pid=61964) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1245.725377] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1245.725377] env[61964]: value = "session[52806a44-4ef2-d4df-9b65-1f773c0a4d7d]52a27453-9094-cb89-1b30-d2ce1a4841d6" [ 1245.725377] env[61964]: _type = "HttpNfcLease" [ 1245.725377] env[61964]: }. {{(pid=61964) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1245.725923] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d1fd6a-344c-4bc1-b50e-9d1f2aeac6a7 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.732761] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk from lease info. {{(pid=61964) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1245.732943] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk. {{(pid=61964) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1245.796267] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-feb967f9-b3f9-41d4-a715-39c4a241a8e2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.869528] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Completed reading data from the image iterator. {{(pid=61964) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1246.869937] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1246.870728] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31a7023-5420-4d54-baac-dad30e8261b9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.877270] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk is in state: ready. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1246.877393] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk. {{(pid=61964) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1246.877626] env[61964]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-05c259dd-e992-4e1a-a29d-4aa5ccfe0482 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.066127] env[61964]: DEBUG oslo_vmware.rw_handles [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52010c34-699c-89db-f3a3-1335a3502cbf/disk-0.vmdk. {{(pid=61964) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1247.066373] env[61964]: INFO nova.virt.vmwareapi.images [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Downloaded image file data eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e [ 1247.067244] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841ec75a-3502-45dc-a27a-e6c1e3389ddf {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.082349] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d897c57-eccc-4548-abcf-415310dd3a30 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.116936] env[61964]: INFO nova.virt.vmwareapi.images [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] The imported VM was unregistered [ 1247.119279] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Caching image {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1247.119523] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Creating directory with path [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1247.119774] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e57287db-ed59-4217-b8fd-d6ad802ef3a1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.129947] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Created directory with path [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e {{(pid=61964) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1247.130160] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219/OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219.vmdk to [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk. {{(pid=61964) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1247.130418] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-aaca374a-2f99-4e7b-a256-09bfd7bcccac {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.137162] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1247.137162] env[61964]: value = "task-1041704" [ 1247.137162] env[61964]: _type = "Task" [ 1247.137162] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.144902] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.646750] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.147961] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.649081] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.149365] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.650340] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041704, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.319896} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.650529] env[61964]: INFO nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219/OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219.vmdk to [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk. [ 1249.650767] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Cleaning up location [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219 {{(pid=61964) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1249.650949] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5f8ead04-b556-45a4-b131-f430c9274219 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1249.651220] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e900a083-5eaf-4cd0-83ac-3e6ea0d0dcb3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.657390] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1249.657390] env[61964]: value = "task-1041705" [ 1249.657390] env[61964]: _type = "Task" [ 1249.657390] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.666134] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.167858] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.057463} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.168228] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1250.168314] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Releasing lock "[datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.168545] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk to [datastore2] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1250.168784] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83943f47-fc90-45ad-ba1e-f8f044db0138 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.174528] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1250.174528] env[61964]: value = "task-1041706" [ 1250.174528] env[61964]: _type = "Task" [ 1250.174528] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.181709] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.684606] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.185376] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.686786] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.187080] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.688692] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041706, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.154153} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.688980] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e/eedcb9f7-7faf-4e3b-ab2f-f9f0a38e427e.vmdk to [datastore2] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk {{(pid=61964) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1252.689769] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee07495-851b-4a3c-858b-a2ef7ada9649 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.710743] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1252.710974] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c4a8a8f-b7e1-4280-8fbd-191f5e195610 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.730249] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1252.730249] env[61964]: value = "task-1041707" [ 1252.730249] env[61964]: _type = "Task" [ 1252.730249] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.737557] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.240458] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.741532] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.242084] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041707, 'name': ReconfigVM_Task, 'duration_secs': 1.399228} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.242527] env[61964]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Reconfigured VM instance instance-0000006e to attach disk [datastore2] a5ee87bf-770f-45be-83d5-16d175102430/a5ee87bf-770f-45be-83d5-16d175102430.vmdk or device None with type streamOptimized {{(pid=61964) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1254.243018] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-464abf92-8a57-44b8-b38d-0e4dfeba4b6a {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.250532] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1254.250532] env[61964]: value = "task-1041708" [ 1254.250532] env[61964]: _type = "Task" [ 1254.250532] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.259112] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041708, 'name': Rename_Task} progress is 5%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.760800] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041708, 'name': Rename_Task, 'duration_secs': 0.145681} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.761048] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powering on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1254.761301] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd8d956b-3f9c-4723-82ea-9a6967e55ac9 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.768254] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1254.768254] env[61964]: value = "task-1041709" [ 1254.768254] env[61964]: _type = "Task" [ 1254.768254] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.775324] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.277795] env[61964]: DEBUG oslo_vmware.api [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041709, 'name': PowerOnVM_Task, 'duration_secs': 0.437589} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.278212] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powered on the VM {{(pid=61964) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1255.376311] env[61964]: DEBUG nova.compute.manager [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Checking state {{(pid=61964) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1255.377283] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80397c8-04a3-49a4-82f8-59eaf9a8cf67 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.897986] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cfe77b30-6a0f-4f0f-8438-c546382ff8a5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.447s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.765997] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.766588] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.766588] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.766079] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.766469] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.766535] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.766706] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.766857] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1259.766852] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.767240] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1259.767240] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 1260.309731] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1260.309894] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquired lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1260.310062] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Forcefully refreshing network info cache for instance {{(pid=61964) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1260.310225] env[61964]: DEBUG nova.objects.instance [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lazy-loading 'info_cache' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1262.039799] env[61964]: DEBUG nova.network.neutron [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [{"id": "4c776c76-6266-46eb-8acb-3896269c2dc5", "address": "fa:16:3e:79:20:8a", "network": {"id": "51e86e44-8bf1-4c70-a1c3-631911962517", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2143652353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d63f8afe09843efb54de6ea85f276aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f3d7e184-c87f-47a5-8d0d-9fa20e07e669", "external-id": "nsx-vlan-transportzone-746", "segmentation_id": 746, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c776c76-62", "ovs_interfaceid": "4c776c76-6266-46eb-8acb-3896269c2dc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.542061] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Releasing lock "refresh_cache-a5ee87bf-770f-45be-83d5-16d175102430" {{(pid=61964) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1262.542312] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updated the network info_cache for instance {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10441}} [ 1263.765793] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.268780] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.269040] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.269232] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.269432] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1264.270360] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e07251-495b-40dc-ac19-bfa156c71bc6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.278424] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c14bb0-1659-4373-86d4-c141f1393a35 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.292535] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c6608b-7f28-459e-8911-f734c22db227 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.298795] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c186c924-8bfa-4985-bdbe-d8177c26c976 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.327673] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181043MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1264.327826] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.328025] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.353045] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Instance a5ee87bf-770f-45be-83d5-16d175102430 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61964) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1265.353045] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1265.353045] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1265.379672] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4f9416-b590-4396-8304-98152e55e8e3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.387612] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d66b272-fa69-4d9f-9321-10995e674dfd {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.416190] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd750434-491e-4635-b95c-162e35547222 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.422782] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b177bff-dc44-4e9f-9acd-8dd3851bd90b {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.436216] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1265.938913] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1266.444297] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1266.444630] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.247921] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.248337] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.248517] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "a5ee87bf-770f-45be-83d5-16d175102430-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1292.248797] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1292.249075] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.252605] env[61964]: INFO nova.compute.manager [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Terminating instance [ 1292.756498] env[61964]: DEBUG nova.compute.manager [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Start destroying the instance on the hypervisor. {{(pid=61964) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1292.756797] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Destroying instance {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1292.757669] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8b481b-a313-46ae-a028-9706c9e1ddd1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.765612] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powering off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1292.765856] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-975e4f64-7c63-4228-afb3-a02bf187ecd6 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.771852] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1292.771852] env[61964]: value = "task-1041710" [ 1292.771852] env[61964]: _type = "Task" [ 1292.771852] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.779847] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041710, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.281089] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041710, 'name': PowerOffVM_Task, 'duration_secs': 0.173014} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.281480] env[61964]: DEBUG nova.virt.vmwareapi.vm_util [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Powered off the VM {{(pid=61964) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1293.281534] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Unregistering the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1293.281785] env[61964]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81167ffc-966f-41b0-afbc-785290233947 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.344420] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Unregistered the VM {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1293.344712] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleting contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1293.344888] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleting the datastore file [datastore2] a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1293.345200] env[61964]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fae93d7a-6c74-4f0b-9272-7d517461e6f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.353010] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for the task: (returnval){ [ 1293.353010] env[61964]: value = "task-1041712" [ 1293.353010] env[61964]: _type = "Task" [ 1293.353010] env[61964]: } to complete. {{(pid=61964) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.360249] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.863133] env[61964]: DEBUG oslo_vmware.api [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Task: {'id': task-1041712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129661} completed successfully. {{(pid=61964) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.863426] env[61964]: DEBUG nova.virt.vmwareapi.ds_util [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted the datastore file {{(pid=61964) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1293.863597] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deleted contents of the VM from datastore datastore2 {{(pid=61964) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1293.863782] env[61964]: DEBUG nova.virt.vmwareapi.vmops [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Instance destroyed {{(pid=61964) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1293.863959] env[61964]: INFO nova.compute.manager [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1293.864235] env[61964]: DEBUG oslo.service.loopingcall [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61964) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1293.864435] env[61964]: DEBUG nova.compute.manager [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Deallocating network for instance {{(pid=61964) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1293.864526] env[61964]: DEBUG nova.network.neutron [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] deallocate_for_instance() {{(pid=61964) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1294.362761] env[61964]: DEBUG nova.compute.manager [req-ba8e8333-d678-4257-8174-bd418c75ac41 req-de484756-583c-47d8-856e-dc35e7e08e11 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Received event network-vif-deleted-4c776c76-6266-46eb-8acb-3896269c2dc5 {{(pid=61964) external_instance_event /opt/stack/nova/nova/compute/manager.py:11573}} [ 1294.363043] env[61964]: INFO nova.compute.manager [req-ba8e8333-d678-4257-8174-bd418c75ac41 req-de484756-583c-47d8-856e-dc35e7e08e11 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Neutron deleted interface 4c776c76-6266-46eb-8acb-3896269c2dc5; detaching it from the instance and deleting it from the info cache [ 1294.363107] env[61964]: DEBUG nova.network.neutron [req-ba8e8333-d678-4257-8174-bd418c75ac41 req-de484756-583c-47d8-856e-dc35e7e08e11 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1294.836354] env[61964]: DEBUG nova.network.neutron [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Updating instance_info_cache with network_info: [] {{(pid=61964) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1294.866028] env[61964]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f14e8af-2506-46ec-a1f3-c5779cf8cd44 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.875250] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d480fc-f9b1-4e99-a54c-21b4f29ae6b1 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.898406] env[61964]: DEBUG nova.compute.manager [req-ba8e8333-d678-4257-8174-bd418c75ac41 req-de484756-583c-47d8-856e-dc35e7e08e11 service nova] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Detach interface failed, port_id=4c776c76-6266-46eb-8acb-3896269c2dc5, reason: Instance a5ee87bf-770f-45be-83d5-16d175102430 could not be found. {{(pid=61964) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11407}} [ 1295.339034] env[61964]: INFO nova.compute.manager [-] [instance: a5ee87bf-770f-45be-83d5-16d175102430] Took 1.47 seconds to deallocate network for instance. [ 1295.845625] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.846181] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.846181] env[61964]: DEBUG nova.objects.instance [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lazy-loading 'resources' on Instance uuid a5ee87bf-770f-45be-83d5-16d175102430 {{(pid=61964) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1296.384696] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071ddcc5-8d9a-42c8-8451-68657a655a28 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.392233] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6b78ed-b735-451b-9b36-8c637d1b5773 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.421260] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a71e4fc-edcb-488b-93d5-36f266823e85 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.428352] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ad397a-5801-430b-996f-17f405d62772 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.442162] env[61964]: DEBUG nova.compute.provider_tree [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1296.945605] env[61964]: DEBUG nova.scheduler.client.report [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1297.450670] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.605s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1297.472353] env[61964]: INFO nova.scheduler.client.report [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Deleted allocations for instance a5ee87bf-770f-45be-83d5-16d175102430 [ 1297.980427] env[61964]: DEBUG oslo_concurrency.lockutils [None req-cce2895d-ad9b-49af-bc79-0fdda044edc5 tempest-AttachVolumeShelveTestJSON-1681313003 tempest-AttachVolumeShelveTestJSON-1681313003-project-member] Lock "a5ee87bf-770f-45be-83d5-16d175102430" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.732s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1320.445248] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.445569] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.445569] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.445647] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.445807] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.765882] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.766166] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.766350] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61964) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10989}} [ 1321.766796] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.767329] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Starting heal instance info cache {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10370}} [ 1321.767329] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Rebuilding the list of instances to heal {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10374}} [ 1322.269716] env[61964]: DEBUG nova.compute.manager [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Didn't find any instances for network info cache update. {{(pid=61964) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10456}} [ 1324.765929] env[61964]: DEBUG oslo_service.periodic_task [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Running periodic task ComputeManager.update_available_resource {{(pid=61964) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.269051] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.269305] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.269481] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1325.269633] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61964) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1325.270611] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dbc72d-e404-42a1-b5da-c66adb2fb989 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.278671] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c310e2a-b5a6-4a02-af5c-45436f149cb2 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.292385] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10aaeeb0-5ab8-4183-91df-58de743dc4a4 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.298354] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc38dab6-09c1-424a-b2e0-c5e94c55e325 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.326963] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181535MB free_disk=183GB free_vcpus=48 pci_devices=None {{(pid=61964) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1325.327107] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1325.327322] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1326.346687] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1326.346964] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61964) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1326.360389] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e763e5f9-c6f6-4a54-b235-e720dcf4ea50 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.368466] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87908281-b16b-46da-a3da-4f2e62277de0 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.397887] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd61380-8002-4df3-b3d1-efbdbf678cde {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.405205] env[61964]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100acbfd-4be5-43c0-85a3-20514e66f1f3 {{(pid=61964) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.418082] env[61964]: DEBUG nova.compute.provider_tree [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed in ProviderTree for provider: 57b292ab-02d9-4aab-ba83-292890345a17 {{(pid=61964) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1326.921441] env[61964]: DEBUG nova.scheduler.client.report [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Inventory has not changed for provider 57b292ab-02d9-4aab-ba83-292890345a17 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 183, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61964) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1327.428271] env[61964]: DEBUG nova.compute.resource_tracker [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61964) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1327.428709] env[61964]: DEBUG oslo_concurrency.lockutils [None req-7f2445b3-3fe7-42d1-ae8f-05623e0f44ba None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.101s {{(pid=61964) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}